Proceedings of the 15th Conference of the European Chapter of The Association for Computational Linguistics: Volume 2 2017
DOI: 10.18653/v1/e17-2113
|View full text |Cite
|
Sign up to set email alerts
|

F-Score Driven Max Margin Neural Network for Named Entity Recognition in Chinese Social Media

Abstract: We focus on named entity recognition (NER) for Chinese social media. With massive unlabeled text and quite limited labelled corpus, we propose a semisupervised learning model based on B-LSTM neural network. To take advantage of traditional methods in NER such as CRF, we combine transition probability with deep learning in our model. To bridge the gap between label accuracy and F-score of NER, we construct a model which can be directly trained on F-score. When considering the instability of Fscore driven method… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
57
1
1

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 107 publications
(59 citation statements)
references
References 19 publications
0
57
1
1
Order By: Relevance
“…One significant drawback is that there is only a very small amount of annotated data available. Weibo NER dataset (Peng and Dredze, 2015;He and Sun, 2017a) and Sighan2006 NER dataset (Levow, 2006) are two widely used datasets for Chinese NER task, containing 1.3k and 45k training examples, respectively. On the two datasets, the highest F1 scores are 48.41% and 89.21%, respectively.…”
Section: Introductionmentioning
confidence: 99%
“…One significant drawback is that there is only a very small amount of annotated data available. Weibo NER dataset (Peng and Dredze, 2015;He and Sun, 2017a) and Sighan2006 NER dataset (Levow, 2006) are two widely used datasets for Chinese NER task, containing 1.3k and 45k training examples, respectively. On the two datasets, the highest F1 scores are 48.41% and 89.21%, respectively.…”
Section: Introductionmentioning
confidence: 99%
“…Peng et al [17] utilize word boundary tags as features to provide richer information and improve the F1-score to 58.99%. He et al [8] propose a unified model for cross domain and improve F1-score to 58.23% from 54.82% [7]. Zhang et al [26] investigate a lattice network which explicitly leverages word and word sequence information, and achieve F1-score of 58.79%.…”
Section: Comparison With Previous Workmentioning
confidence: 99%
“…The model (Peng and Dredze, 2016) that jointly trains NER and CWS reaches a F1-score of 58.99%. He and Sun (2017b) propose a unified model to exploit crossdomain and semi-supervised data, which improves the F1-score from 54.82% to 58.23% compared with the model proposed by He and Sun (2017a). Cao et al (2018) use an adversarial transfer learning framework to incorporate task-shared word boundary information from CWS and achieves a F1-score of 58.70%.…”
Section: Weibo Datasetmentioning
confidence: 99%