Proceedings of the Conference on Empirical Methods in Natural Language Processing - EMNLP '08 2008
DOI: 10.3115/1613715.1613751
|View full text |Cite
|
Sign up to set email alerts
|

Cheap and fast---but is it good?

Abstract: Human linguistic annotation is crucial for many natural language processing tasks but can be expensive and time-consuming. We explore the use of Amazon's Mechanical Turk system, a significantly cheaper and faster method for collecting annotations from a broad base of paid non-expert contributors over the Web. We investigate five tasks: affect recognition, word similarity, recognizing textual entailment, event temporal ordering, and word sense disambiguation. For all five, we show high agreement between Mechani… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
141
0
1

Year Published

2010
2010
2016
2016

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 1,022 publications
(142 citation statements)
references
References 24 publications
0
141
0
1
Order By: Relevance
“…To obtain labeled training data for the classifier, we utilized workers from the Amazon Mechanical Turk 7 . It has been shown that manual labeling from Amazon Turk can correlate well with experts [11]. We used thousands of workers to assign sentiments for a large random sample of tweets, ensuring that each tweet was labeled by three different people.…”
Section: Sentiment Analysismentioning
confidence: 99%
“…To obtain labeled training data for the classifier, we utilized workers from the Amazon Mechanical Turk 7 . It has been shown that manual labeling from Amazon Turk can correlate well with experts [11]. We used thousands of workers to assign sentiments for a large random sample of tweets, ensuring that each tweet was labeled by three different people.…”
Section: Sentiment Analysismentioning
confidence: 99%
“…In [9], a high agreement between MTurk non-expert annotations and existing gold-standard labels provided by expert labelers for five natural language processing tasks is demonstrated. Multiple labeling has also been shown to be useful for improving the data quality of annotations of non-experts [5].…”
Section: Related Workmentioning
confidence: 88%
“…Earlier work on inferring ground truth from subjective labels includes Smyth et al [6], Sheng et al [5], and Snow et al [9]. In [6], the latent relation between subjective labels and true labels by EM algorithm was studied and it was shown that the posterior conditional probabilities of subjective labels and true labels generally agree with intuition and often (70% of the samples) correspond to a majority vote among the labelers.…”
Section: Related Workmentioning
confidence: 99%
“…With the increasing popularity, a set of crowdsourcing market platforms such as Amazon Mechanical Turk and CrowdFlower [5] have emerged, which enable human workers to perform tasks on the Internet. Crowdsourcing applications have been adopted in a wide range of applications such as image search [24], natural language annotations [22] and information retrieval [11]. Moreover, crowdsourcing has also been incorporated into database design and relational query processing [10,12,18].…”
Section: Related Workmentioning
confidence: 99%