Proceedings of the 1st Workshop on Multilingual Representation Learning 2021
DOI: 10.18653/v1/2021.mrl-1.19
|View full text |Cite
|
Sign up to set email alerts
|

Analysis of Zero-Shot Crosslingual Learning between English and Korean for Named Entity Recognition

Abstract: This paper presents a English-Korean parallel dataset that collects 381K news articles where 1,400 of them, comprising 10K sentences, are manually labeled for crosslingual named entity recognition (NER). The annotation guidelines for the two languages are developed in parallel, that yield the inter-annotator agreement scores of 91 and 88% for English and Korean respectively, indicating sublime quality annotation in our dataset. Three types of crosslingual learning approaches, direct model transfer, embedding p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 22 publications
0
0
0
Order By: Relevance
“…The information bottleneck (IB) was first proposed in [30]. The principle of IB is to learn a good representation that can retain task-relevance information while reducing taskirrelevance information [17,2]. Since IB is plagued by computational mutual information, many methods are devoted to finding its lower bound.…”
Section: Related Workmentioning
confidence: 99%
“…The information bottleneck (IB) was first proposed in [30]. The principle of IB is to learn a good representation that can retain task-relevance information while reducing taskirrelevance information [17,2]. Since IB is plagued by computational mutual information, many methods are devoted to finding its lower bound.…”
Section: Related Workmentioning
confidence: 99%