2021
DOI: 10.1145/3476076
|View full text |Cite
|
Sign up to set email alerts
|

Goldilocks: Consistent Crowdsourced Scalar Annotations with Relative Uncertainty

Abstract: Human ratings have become a crucial resource for training and evaluating machine learning systems. However, traditional elicitation methods for absolute and comparative rating suffer from issues with consistency and often do not distinguish between uncertainty due to disagreement between annotators and ambiguity inherent to the item being rated. In this work, we present Goldilocks, a novel crowd rating elicitation technique for collecting calibrated scalar annotations that also distinguishes inherent ambiguity… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 62 publications
0
2
0
Order By: Relevance
“…Some researchers propose the use of rationales as answers (Donahue and Grauman 2011;McDonnell et al 2016), while others propose open-ended answers that are then clustered or taxonomized (Kairam and Heer 2016;Chang, Amershi, and Kamar 2017). Finally, some have proposed more middle ground solutions in the form of new annotation representations, such as ranges in scalar rating annotation (Chen, Weld, and Zhang 2021). Instead of asking for confidence or uncertainty via a question separate from the annotation itself, range annotations enable annotators to directly convey uncertainty calibrated to their annotation.…”
Section: Capturing Uncertainty In Annotationmentioning
confidence: 99%
“…Some researchers propose the use of rationales as answers (Donahue and Grauman 2011;McDonnell et al 2016), while others propose open-ended answers that are then clustered or taxonomized (Kairam and Heer 2016;Chang, Amershi, and Kamar 2017). Finally, some have proposed more middle ground solutions in the form of new annotation representations, such as ranges in scalar rating annotation (Chen, Weld, and Zhang 2021). Instead of asking for confidence or uncertainty via a question separate from the annotation itself, range annotations enable annotators to directly convey uncertainty calibrated to their annotation.…”
Section: Capturing Uncertainty In Annotationmentioning
confidence: 99%
“…3.5.2 Providing insights from past annotation decisions. Exploring different strategies for better grasping the importance of elements in the annotation task, possibly by relying on the collective wisdom or opinions of others, has been useful in past literature[17,18]. P4 mentions sharing what other people thought or getting insights from others' perspectives might be more helpful in determining the significance of certain aspects of the task, as he remarked, "This was relevant because of this aspect or something like that.…”
mentioning
confidence: 99%