Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.637
|View full text |Cite
|
Sign up to set email alerts
|

Cold-start Active Learning through Self-supervised Language Modeling

Abstract: Active learning strives to reduce annotation costs by choosing the most critical examples to label. Typically, the active learning strategy is contingent on the classification model. For instance, uncertainty sampling depends on poorly calibrated model confidence scores. In the cold-start setting, active learning is impractical because of model instability and data scarcity. Fortunately, modern NLP provides an additional source of information: pretrained language models. The pre-training loss can find examples… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
93
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4
2

Relationship

1
9

Authors

Journals

citations
Cited by 90 publications
(93 citation statements)
references
References 39 publications
0
93
0
Order By: Relevance
“…Ruiter et al (2019) use an emergent NMT system to simultaneously select training data and learn internal NMT representations in a SSL way without parallel data. SSL is also adopted to solve many other problems, such as document-level context or sentence summarization (West et al, 2019;Wang et al, 2019b), dialogue learning (Wu et al, 2019), improving data scarcity or labeling costs (Fu et al, 2020;Yuan et al, 2020) and generating meta-learning tasks from unlabeled text (Bansal et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Ruiter et al (2019) use an emergent NMT system to simultaneously select training data and learn internal NMT representations in a SSL way without parallel data. SSL is also adopted to solve many other problems, such as document-level context or sentence summarization (West et al, 2019;Wang et al, 2019b), dialogue learning (Wu et al, 2019), improving data scarcity or labeling costs (Fu et al, 2020;Yuan et al, 2020) and generating meta-learning tasks from unlabeled text (Bansal et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Ideally, active learning could be most useful in low-resource settings. In practice, however, it is more likely that the model might work poorly with the limited number of labelled data at the beginning of active learning [34]. Therefore, introducing a component to ensures a certain level of performance with the limited labelled data is important to address the cold-start issue.…”
Section: Related Work a Aspect-based Sentiment Analysismentioning
confidence: 99%
“…While general purpose transformer-based language models are often used, fine-tuning them often require numerous examples (Dodge et al, 2020). One possible solution to address this lack of labelled data is to use active learning (Yuan et al, 2020) in conjunction with pre-trained models such as BERT (Devlin et al, 2019) that comes with possible bias problems (Papakyriakopoulos et al, 2020). In Cultural Heritage domains, several works are trying to compensate for the lack of labelled data.…”
Section: Related Workmentioning
confidence: 99%