Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.611
|View full text |Cite
|
Sign up to set email alerts
|

FLAT: Chinese NER Using Flat-Lattice Transformer

Abstract: Recently, the character-word lattice structure has been proved to be effective for Chinese named entity recognition (NER) by incorporating the word information. However, since the lattice structure is complex and dynamic, most existing lattice-based models are hard to fully utilize the parallel computation of GPUs and usually have a low inference-speed. In this paper, we propose FLAT: Flat-LAttice Transformer for Chinese NER, which converts the lattice structure into a flat structure consisting of spans. Each … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
163
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

2
7

Authors

Journals

citations
Cited by 343 publications
(163 citation statements)
references
References 19 publications
0
163
0
Order By: Relevance
“…Last but not least, the pre-trained language model BERT has been extensively exploited in the Natural Language Processing (NLP) community since its introduction (Devlin et al, 2019;Conneau and Lample, 2019). Owing to BERT's ability to extract contextualized information, BERT has been successfully utilized to enhance various tasks substantially, such as the aspect-based sentiment analysis task , summarization (Zhong et al, 2019), named entity recognition (Yan et al, 2019;Li et al, 2020) and Chinese dependency parsing . However, most works used BERT as an encoder, and less work uses BERT to do generation (Wang and Cho, 2019;Conneau and Lample, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…Last but not least, the pre-trained language model BERT has been extensively exploited in the Natural Language Processing (NLP) community since its introduction (Devlin et al, 2019;Conneau and Lample, 2019). Owing to BERT's ability to extract contextualized information, BERT has been successfully utilized to enhance various tasks substantially, such as the aspect-based sentiment analysis task , summarization (Zhong et al, 2019), named entity recognition (Yan et al, 2019;Li et al, 2020) and Chinese dependency parsing . However, most works used BERT as an encoder, and less work uses BERT to do generation (Wang and Cho, 2019;Conneau and Lample, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…It uses the existing neural network model to model the input sequence. The main neural network models are network models based on RNN and its variants [42,43], models based on CNN [44,45], and models based on Transformer [46,47].…”
Section: Model Frameworkmentioning
confidence: 99%
“…The network model based on Transformer. Li Xiaonan et al [46] proposed a FLAT(Flat-LAttice Transformer) structure to be applied to Chinese NER. This model relies on the powerful functions of Transformer and carefully designed specific location codes to fully utilize lattice information and has efficient parallelism.…”
Section: Model Frameworkmentioning
confidence: 99%
“…A high quality of text representation plays an important role to obtain good performance for many NLP tasks (Song et al, 2017;Zhu et al, 2019;Liu and Lapata, 2019), where a powerful encoder is required to model more contextual information. Inspired by the studies (Song et al, 2009;Song and Xia, 2012;Ouyang et al, 2017;Kim et al, 2018;Peng et al, 2018;Higashiyama et al, 2019;Tian et al, 2020c;Li et al, 2020) that leverage the large granularity contextual information carried by n-grams to enhance text representation for Chinese, we propose ZEN to enhance character based text encoders (e.g., BERT) by leveraging ngrams. In doing so, we extract n-grams prior to pre-training ZEN through two different steps.…”
Section: N-gram Extractionmentioning
confidence: 99%