The World Wide Web Conference 2019
DOI: 10.1145/3308558.3313743
|View full text |Cite
|
Sign up to set email alerts
|

Neural Chinese Named Entity Recognition via CNN-LSTM-CRF and Joint Training with Word Segmentation

Abstract: Chinese named entity recognition (CNER) is an important task in Chinese natural language processing field. However, CNER is very challenging since Chinese entity names are highly contextdependent. In addition, Chinese texts lack delimiters to separate words, making it difficult to identify the boundary of entities. Besides, the training data for CNER in many domains is usually insufficient, and annotating enough training data for CNER is very expensive and time-consuming. In this paper, we propose a neural app… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
27
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 78 publications
(27 citation statements)
references
References 26 publications
0
27
0
Order By: Relevance
“…Some other hybrid approaches have also reported significantly enhanced performances of NER by combining LSTM and CNN (e.g. [29][30][31][32][33][34]). To name a few, Chiu [29] implemented CNNs to identify the character-level features and then exploited BiLSTM-based modules for sequence-labelling.…”
Section: B Dl-based Neural Network For Ner Tasksmentioning
confidence: 99%
See 2 more Smart Citations
“…Some other hybrid approaches have also reported significantly enhanced performances of NER by combining LSTM and CNN (e.g. [29][30][31][32][33][34]). To name a few, Chiu [29] implemented CNNs to identify the character-level features and then exploited BiLSTM-based modules for sequence-labelling.…”
Section: B Dl-based Neural Network For Ner Tasksmentioning
confidence: 99%
“…For the second phase, we propose an encoder-decoder architecture for NER sequential tagger by adapting the existing model structure of "LSTM-CNNs-CRF" [40,41], as illustrated in the right half of Figure 1. This advanced model includes four layers: an embedding layer, a local context layer, a sequential semantic layer and a generative layer.…”
Section: Proposed Frameworkmentioning
confidence: 99%
See 1 more Smart Citation
“…For example, the English word segmentation is generally divided into words based on spaces, while Chinese is more complicated. There is generally no obvious division mark between words, and different methods need to be selected according to different task needs [25]. A programming language is similar to natural language and can also be viewed as a sequence of English characters.…”
Section: ) Function To Tokensmentioning
confidence: 99%
“…However, the pipeline method suffers from error propagation, since the error of CWS may affect the performance of NER. The second one is to learn CWS and NER tasks jointly (Xu et al, 2013;Peng and Dredze, 2016;Cao et al, 2018;Wu et al, 2019). However, the joint models must rely on CWS annotation datasets, which are costly and are annotated under many diverse segmentation criteria (Chen et al, 2017).…”
Section: Introductionmentioning
confidence: 99%