2021
DOI: 10.1109/taslp.2021.3073868
|View full text |Cite
|
Sign up to set email alerts
|

FSPRM: A Feature Subsequence Based Probability Representation Model for Chinese Word Embedding

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 36 publications
0
3
0
Order By: Relevance
“…Strokes are the most fine-grained units in Chinese characters, and each Chinese character can be decomposed into a sequence of strokes with a specific order. According to the inspiration of [29], we categorize strokes into five types: horizontal, vertical, apostrophe, dot, and fold, which are also the five basic types of strokes stipulated in the "Common Character List of Modern Chinese" jointly issued by several Chinese cultural departments in 1988. We coded these 5 types of strokes as shown in Table 2.…”
Section: Methodsmentioning
confidence: 99%
“…Strokes are the most fine-grained units in Chinese characters, and each Chinese character can be decomposed into a sequence of strokes with a specific order. According to the inspiration of [29], we categorize strokes into five types: horizontal, vertical, apostrophe, dot, and fold, which are also the five basic types of strokes stipulated in the "Common Character List of Modern Chinese" jointly issued by several Chinese cultural departments in 1988. We coded these 5 types of strokes as shown in Table 2.…”
Section: Methodsmentioning
confidence: 99%
“… Chuang et al ( 2021 ) Automatic speech recognition and translation system LibriSpeech corpus, Augmented LibriSpeech, Fisher Spanish corpora LSTM, BiLSTM, CNN Word2Vec, fastText Word2Vec model efficiently maps speech signals to semantic space 7. Zhang et al ( 2021 ) Chinese word representation SogouCA data, Wikipedia dump, Fudan dataset LSTM Word2Vec, GloVe, BERT, CWE LSTM + CWE achieves an F1-score of 95.53% for the NER task 8. Shekhar et al ( 2019 ) English-Hindi mixed languages text identification Dataset ICON 2016, Forum for IR Evaluation 2014 shared task on transliterated search, MSIR 2015, 2016 BiLSTM Word2Vec, Character and word embedding BiLSTM + Word2Vec achieves an F1-score of 83.9% for the NER task 9.…”
Section: Appendix Amentioning
confidence: 99%
“…The Chinese word embedding-based model with LSTM acquires an F1-score of 95.53% to understand the semantics of words and efficiently analyze the features. (Zhang et al 2021). A domain-specific word embedding approach with a fuzzy metric that focuses on a unique entity recognition task is proposed to adopt cooking recipes from a set of all available recipes.…”
Section: Named Entity Recognition and Recommendation Systemmentioning
confidence: 99%