Proceedings of the 55th Annual Meeting of the Association For Computational Linguistics (Volume 1: Long Papers) 2017
DOI: 10.18653/v1/p17-1135
|View full text |Cite
|
Sign up to set email alerts
|

Weakly Supervised Cross-Lingual Named Entity Recognition via Effective Annotation and Representation Projection

Abstract: The state-of-the-art named entity recognition (NER) systems are supervised machine learning models that require large amounts of manually annotated data to achieve high accuracy. However, annotating NER data by human is expensive and time-consuming, and can be quite difficult for a new language. In this paper, we present two weakly supervised approaches for cross-lingual NER with no human annotation in a target language. The first approach is to create automatically labeled NER data for a target language via a… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
130
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 123 publications
(131 citation statements)
references
References 21 publications
1
130
0
Order By: Relevance
“…Regard- less, the BERT zero-resource performance far exceeds the results published in previous work. Mayhew et al (2017) and Ni et al (2017) do use some cross-lingual resources (like bilingual dictionaries) in their experiments, but it appears that BERT with multilingual pretraining performs better, even though it does not have access to crosslingual information. Table 3: Median cosine similarity between the mean-pooled BERT embeddings of MLDoc English documents and their translations, with and without language-adversarial training.…”
Section: Conll Ner Resultsmentioning
confidence: 97%
“…Regard- less, the BERT zero-resource performance far exceeds the results published in previous work. Mayhew et al (2017) and Ni et al (2017) do use some cross-lingual resources (like bilingual dictionaries) in their experiments, but it appears that BERT with multilingual pretraining performs better, even though it does not have access to crosslingual information. Table 3: Median cosine similarity between the mean-pooled BERT embeddings of MLDoc English documents and their translations, with and without language-adversarial training.…”
Section: Conll Ner Resultsmentioning
confidence: 97%
“…This requires first projecting annotations from the source data to the (unlabeled) target data. Many approaches in this category rely upon parallel corpora (Yarowsky et al, 2001;Zeman and Resnik, 2008;Ehrmann et al, 2011;Fu et al, 2011;Ni et al, 2017), first annotating the source data using a trained model and then projecting the annotations. Only a few works explore the use of MT to first translate a gold annotated corpus to obtain a synthetic parallel corpus and then project annotations (Tiedemann et al, 2014).…”
Section: Annotation Projectionmentioning
confidence: 99%
“…When projecting annotations, one encounters the problem of word alignment. Most of the existing works (Yarowsky et al, 2001;Shah et al, 2010;Ni et al, 2017) rely upon unsupervised alignment models from statistical MT literature, such as IBM Models 1-6 (Brown et al, 1993;Och and Ney, 2003). Other works focus on low-resource settings (Mayhew et al, 2017;Xie et al, 2018) perform translation word-by-word or phrase-by-phrase, and thus do not need to perform word alignment.…”
Section: Annotation Projectionmentioning
confidence: 99%
See 1 more Smart Citation
“…In annotation projection approaches, parallel or comparable corpora are commonly used (Yarowsky et al, 2001;Ehrmann et al, 2011;Das and Petrov, 2011;Li et al, 2012;Täckström et al, 2013;Wang and Manning, 2014;Ni et al, 2017). The source language sentences of parallel corpora are first annotated either manually or by a pretrained tagger.…”
Section: Annotation Projectionmentioning
confidence: 99%