2021
DOI: 10.48550/arxiv.2112.07475
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Two Contrasting Data Annotation Paradigms for Subjective NLP Tasks

Abstract: Labelled data is the foundation of most natural language processing tasks. However, labelling data is difficult and there often are diverse valid beliefs about what the correct data labels should be. So far, dataset creators have acknowledged annotator subjectivity, but not actively managed it in the annotation process. This has led to partly-subjective datasets that fail to serve a clear downstream use. To address this issue, we propose two contrasting paradigms for data annotation. The descriptive paradigm e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 9 publications
(10 citation statements)
references
References 24 publications
0
10
0
Order By: Relevance
“…For each language we recruited at least 12 annotators, so that in total more than 120 annotators worked on MHC. 5 Annotation was prescriptive (Röttger et al, 2021a), meaning that annotators were tasked with adhering to clear annotation guidelines reflecting our definition of hate speech, rather than sharing their personal view on what is and is not hateful. 6 Compared to the original HATECHECK, where four out of five annotators confirmed the gold label for 99.4% of all test cases, there was more disagreement on MHC (see Appendix D).…”
Section: Generating Test Casesmentioning
confidence: 99%
“…For each language we recruited at least 12 annotators, so that in total more than 120 annotators worked on MHC. 5 Annotation was prescriptive (Röttger et al, 2021a), meaning that annotators were tasked with adhering to clear annotation guidelines reflecting our definition of hate speech, rather than sharing their personal view on what is and is not hateful. 6 Compared to the original HATECHECK, where four out of five annotators confirmed the gold label for 99.4% of all test cases, there was more disagreement on MHC (see Appendix D).…”
Section: Generating Test Casesmentioning
confidence: 99%
“…To validate the gold-standard labels assigned to each test case, we recruited three annotators with prior experience on hate speech projects. 5 Annotators were given extensive and prescriptive guidelines (Röttger et al, 2022), as well as test tasks and training sessions, which included examining real-world examples of emoji-based hate from Twitter. We followed guidance for protecting annotator well-being (Vidgen et al, 2019).…”
Section: Validating Test Casesmentioning
confidence: 99%
“…One factor in settling how annotations can support the purpose of the task is how much subjectivity is desired. Röttger et al (2021a) distinguish two types of approaches to annotation: descriptive and prescriptive. Descriptive annotations encourage subjectivity of annotators (where inconsistency is not an issue), while prescription instructs annotators to strictly follow carefully defined criteria (the less subjectivity, the better).…”
Section: Task Specific Annotationsmentioning
confidence: 99%