Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.206
|View full text |Cite
|
Sign up to set email alerts
|

Automated Concatenation of Embeddings for Structured Prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
37
0
2

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 78 publications
(39 citation statements)
references
References 79 publications
(60 reference statements)
0
37
0
2
Order By: Relevance
“…Moreover, the embeddings are trained over long documents, which allows the model to easily model long-range dependencies to disambiguate complex named entities in the sentence. Recently, a lot of work shows that utilizing the document-level contexts in the CoNLL NER datasets can significantly improve token representations and achieves state-of-the-art performance (Yu et al, 2020;Luoma and Pyysalo, 2020;Yamada et al, 2020;Wang et al, 2021a). However, the lack of context in the MultiCoNER datasets means the embeddings cannot take advantage of long-range dependencies for entity disambiguation.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Moreover, the embeddings are trained over long documents, which allows the model to easily model long-range dependencies to disambiguate complex named entities in the sentence. Recently, a lot of work shows that utilizing the document-level contexts in the CoNLL NER datasets can significantly improve token representations and achieves state-of-the-art performance (Yu et al, 2020;Luoma and Pyysalo, 2020;Yamada et al, 2020;Wang et al, 2021a). However, the lack of context in the MultiCoNER datasets means the embeddings cannot take advantage of long-range dependencies for entity disambiguation.…”
Section: Related Workmentioning
confidence: 99%
“…Shi and Lee (2021) proposed two-stage fine-tuning, which first trains a general multilingual Enhanced Universal Dependency (Bouma et al, 2021) parser and then finetunes on each specific language separately. Wang et al (2021a) proposed to train models through concatenating fine-tuned embeddings. We extend these ideas as multi-stage fine-tuning, which improves the accuracy of monolingual models that use finetuned multilingual embeddings as initialization in training.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…While named entity recognition in text has been studied extensively in the NLP community (Mikheev et al, 1999;Florian et al, 2003;Nadeau and Sekine, 2007;Ratinov and Roth, 2009;Ritter et al, 2011;Lample et al, 2016;Chiu and Nichols, 2016;Akbik et al, 2019;Wang et al, 2021b;Yamada et al, 2020), relatively little work has been conducted on extracting named entities from speech (Kim and Woodland, 2000;Sudoh et al, 2006;Parada et al, 2011;Caubrière et al, 2020;Yadav et al, 2020;Shon et al, 2021). Recognizing named entities from speech is a more challenging task which is commonly done through a pipeline approach: combining an automatic speech recognition (ASR) system with a text-based NER model (Sudoh et al, 2006;Raymond, 2013;Jannet et al, 2015).…”
Section: Spoken Named Entity Recognitionmentioning
confidence: 99%
“…Standard Named Entity Recognition (NER) for high resource domains has already been successfully addressed [1,2]. In contrast, NER taggers often cannot achieve satisfying results in the historical domain.…”
Section: Introductionmentioning
confidence: 99%