2018
DOI: 10.48550/arxiv.1805.06504
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Analogical Reasoning on Chinese Morphological and Semantic Relations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
28
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 26 publications
(28 citation statements)
references
References 9 publications
0
28
0
Order By: Relevance
“…Main results on the text dataset THUNEWS-LT We choose a commonly-used subset of Chinese news dataset THUCNews (Sun et al 2016) which contains ten classes 1 . We build its long-tailed version with the imbalance factor at 100 and use the pre-trained embedding model provided by Li et al (2018). We choose three models including TextCNN (Kalchbrenner, Grefenstette, and Blunsom 2014), TextRNN (Liu, Qiu, and Huang 2016) and Transformer (Vaswani et al 2017) as the backbone.…”
Section: Main Results On Inaturalist Tablementioning
confidence: 99%
“…Main results on the text dataset THUNEWS-LT We choose a commonly-used subset of Chinese news dataset THUCNews (Sun et al 2016) which contains ten classes 1 . We build its long-tailed version with the imbalance factor at 100 and use the pre-trained embedding model provided by Li et al (2018). We choose three models including TextCNN (Kalchbrenner, Grefenstette, and Blunsom 2014), TextRNN (Liu, Qiu, and Huang 2016) and Transformer (Vaswani et al 2017) as the backbone.…”
Section: Main Results On Inaturalist Tablementioning
confidence: 99%
“…The collection dataset is obtained on a MOOC platform with complex data dimensions. Therefore, a large Chinese corpus is used to capture the semantic information between different words [28]. Besides, we use the Word2Vec tool to convert the text corpus as input and convert the title and course profile content into a 300-dimensional feature vector.…”
Section: Text Feature Extractionmentioning
confidence: 99%
“…Perhaps because of the size difference between the two corpuses, increasingly researchers developing cutting edge Chinese langauge NLP models are drawing on the Baidu Baike corpus [38,43]. Baidu Baike word embeddings have been shown to perform better on certain tasks [21]. Here, we assess the downstream implications of this choice on the representation of democratic concepts, social control, and historical events and figures.…”
Section: Censorship Of Chinese Language Wikipedia and Implications Fo...mentioning
confidence: 99%
“…In this section, we consider the differences in word associations among word embeddings trained with Chinese language Wikipedia and Baidu Baike. We use word embeddings made available by Li et al [21]. 4 Li et al [21] train 300-dimensional word embeddings on both Baidu Baike and Chinese language Wikipedia using the same algorithm, Word2Vec [24].…”
Section: Distance From Democracy: Comparison Between Baidu Baike and ...mentioning
confidence: 99%