Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.256
|View full text |Cite
|
Sign up to set email alerts
|

Glyph2Vec: Learning Chinese Out-of-Vocabulary Word Embedding from Glyphs

Abstract: Chinese NLP applications that rely on large text often contain huge amounts of vocabulary which are sparse in corpus. We show that characters' written form, Glyphs, in ideographic languages could carry rich semantics. We present a multi-modal model, Glyph2Vec, to tackle Chinese out-of-vocabulary word embedding problem. Glyph2Vec extracts visual features from word glyphs to expand current word embedding space for out-of-vocabulary word embedding, without the need of accessing any corpus, which is useful for imp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
8
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(8 citation statements)
references
References 18 publications
(27 reference statements)
0
8
0
Order By: Relevance
“…For completeness, we also replace the SVM classifier with a neural network classifier (Chen et al, 2020a) consisting of 3 fully-connected layers. The results are shown in Appendix I.…”
Section: Word Embedding Learningmentioning
confidence: 99%
“…For completeness, we also replace the SVM classifier with a neural network classifier (Chen et al, 2020a) consisting of 3 fully-connected layers. The results are shown in Appendix I.…”
Section: Word Embedding Learningmentioning
confidence: 99%
“…However, from the perspective of symbolic evolution, Chinese symbols have always possessed their unique structural features and peculiarities, developing from the initial hieroglyphics to their present forms as the Chinese characters shown in Figure 1. Recent researches [6,38] also demonstrate that the glyphs of Chinese characters contain rich semantic information and have the potential to enhance the word representation of them. Meng et al [26] first apply the glyph features of Chinese characters into the pre-trained model BERT and achieve significant performance on many Chinese NLU tasks, such as Named Entity Recognition [20], News Text Classification [22] and Sentiment Analysis [40].…”
Section: Related Workmentioning
confidence: 99%
“…Hence, the glyphs of Chinese characters can convey some meanings in many cases, and Chinese characters with similar structures can have intrinsic links. They Based on the above observations, some methods [6,26,38] incorporate the glyph features to enhance the Chinese character representation already covered into character embeddings (character ID-based), e.g., Meng et al [26] combine the glyph features extracted from various forms of Chinese characters with the BERT embeddings. They demonstrate that the glyph features of Chinese characters are authentically helpful to improve the performance of models.…”
Section: Introductionmentioning
confidence: 99%
“…Several related works have been published to explore more feasibility. For instance, GWE directly extracted character glyph features from the bitmaps of characters by convolutional auto-encoder (ConvAE) [27], Glyph2Vec extracted visual features from word glyphs to expand current word embedding space for out-of-vocabulary word embedding, without the need of accessing any corpus [28], and Glyce designed Convolutional Neural Network (CNN) structures (called tianzige-CNN) tailored to Chinese character image processing [29]. These models extract the morphological feature by traditional image processing methods instead of linguistic knowledge, which introduces the connection of characters, and the noise from the unclarity of image processing algorithms.…”
Section: Introductionmentioning
confidence: 99%
“…The results of the word similarity task. ± 0.33 59 28. ± 0.21 58.08 ± 0.32 31.84 ± 0.24 50.77 ± 0.05 57.32 ± 0.28 38.35 ± 0.27 31.93 ± 0.05 Skipgram 58.32 ± 0.40 60.07 ± 0.21 49.18 ± 0.05 33.18 ± 0.22 56.42 ± 0.27 61.6 ± 0.03 31.66 ± 0.11 31.79 ± 0.32 CWE 57.58 ± 0.34 59.46 ± 0.18 50.70 ± 0.01 33.98 ± 0.49 56.76 ± 0.01 62.18 ± 0.21 33.91 ± 0.48 30.84 ± 0.21 JWE 52.04 ± 0.21 62.51 ± 0.68 57.46 ± 0.54 35.06 ± 0.58 54.83 ± 0.15 66.05 ±0.40 43.95 ± 0.09 36.97 ± 0.04 cw2vec 53.81 ± 0.03 57.54 ± 0.30 57.96 ± 0.29 23.88 ± 0.21 56.32 ± 0.07 61.29 ± 0.85 37.61 ± 0.49 27.66 ± 0.18 RECWE 53.73 ± 0.22 61.71 ± 0.14 32.40 ± 0.14 32.24 ± 0.07 53.73 ± 0.22 62.21 ± 0.36 54.4 ± 0.14 32.24 ± 0.07 4CWE 53.18 ± 0.05 61.61 ± 0.46 63.30 ± 0.20 37.47 ± 0.34 53.14 ± 0.21 63.54 ± 0.65 50.38 ± 0.60 39.56 ± 0.14 4CWE(ATW) 53.26 ± 0.30 62.04 ± 0.04 63.36 ±0.25 37.11 ± 0.02 53.19 ± 0.48 63.98 ± 0.33 51.51 ±0.52 39.43 ± 0.03 4CWE(ACW) 53.70 ± 0.38 62.58 ±0.06 60.52 ± 0.94 38.44 ±0.35 53.79 ± 0.15 62.91 ± 0.52 48.82 ± 0.12 40.02 ±0.05…”
mentioning
confidence: 97%