Proceedings of the Third Workshop on Deep Learning for Low-Resource Natural Language Processing 2022
DOI: 10.18653/v1/2022.deeplo-1.17
|View full text |Cite
|
Sign up to set email alerts
|

Clean or Annotate: How to Spend a Limited Data Collection Budget

Abstract: Crowdsourcing platforms are often used to collect datasets for training machine learning models, despite higher levels of inaccurate labeling compared to expert labeling. There are two common strategies to manage the impact of such noise: The first involves aggregating redundant annotations, but comes at the expense of labeling substantially fewer examples. Secondly, prior works have also considered using the entire annotation budget to label as many examples as possible and subsequently apply denoising algori… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 28 publications
0
3
0
Order By: Relevance
“…Tables 3 and 4 report accuracy, precision and recall scores for the 6 classification tasks relative to detecting the presence or absence of a given emotion in a post from GoEmotions 34 . Recent work 71 showed that the original labelling from GoEmotions, over 27 emotional states, contained several invalid annotations. Since there is no evidence that the errors are systematically biased to favour one method over the others, we preserved the original labelling but considered only those labels equivalent to our emotions rather than mapping all 27 emotional states into the emotions supported by BERT (as done in 34 ) or EmoAtlas.…”
Section: Emoatlas and Goemotionsmentioning
confidence: 99%
“…Tables 3 and 4 report accuracy, precision and recall scores for the 6 classification tasks relative to detecting the presence or absence of a given emotion in a post from GoEmotions 34 . Recent work 71 showed that the original labelling from GoEmotions, over 27 emotional states, contained several invalid annotations. Since there is no evidence that the errors are systematically biased to favour one method over the others, we preserved the original labelling but considered only those labels equivalent to our emotions rather than mapping all 27 emotional states into the emotions supported by BERT (as done in 34 ) or EmoAtlas.…”
Section: Emoatlas and Goemotionsmentioning
confidence: 99%
“…Tables 4, 5 and 6 report accuracy, precision and recall scores for the 6 classification tasks relative to detecting the presence or absence of a given emotion in a post from GoEmotions [18]. Recent work [12] showed that the original labelling from GoEmotions, over 27 emotional states, contained several invalid annotations. Since there is no evidence that the errors are systematically biased to favour one method over the others, we preserved the original labelling but considered only those labels equivalent to our emotions rather than mapping all 27 emotional states into the emotions supported by BERT, as done in [18], or EmoAtlas.…”
Section: Emoatlas and Goemotionsmentioning
confidence: 99%
“…Another study by (Zhang et al, 2021) explored new annotation distribution schemes, assigning multiple labels per example for a small subset of training examples, and proposed a learning algorithm that efficiently combines signals from uneven training data. Finally, a study by (Chen et al, 2022) proposed an approach that reserves a fraction of annotations to explicitly clean up highly probable error samples to optimize the annotation process. All these studies contribute to the understanding of how to maximize the performance of NLP models under restricted annotation budgets.…”
Section: Related Workmentioning
confidence: 99%