Proceedings of the 26th Annual International Conference on Machine Learning 2009
DOI: 10.1145/1553374.1553488
|View full text |Cite
|
Sign up to set email alerts
|

Supervised learning from multiple experts

Abstract: We describe a probabilistic approach for supervised learning when we have multiple experts/annotators providing (possibly noisy) labels but no absolute gold standard. The proposed algorithm evaluates the different experts and also gives an estimate of the actual hidden labels. Experimental results indicate that the proposed method is superior to the commonly used majority voting baseline.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
131
0

Year Published

2015
2015
2019
2019

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 272 publications
(131 citation statements)
references
References 10 publications
0
131
0
Order By: Relevance
“…In addition, BCLAs was compared to BCLA, 'smart fusion' (i.e. the benchmarking algorithm proposed by authors of the dataset (Karlen et al 2013)), the bestperforming (lowest-MAE) single algorithm (denoted 'Best'), the EM algorithm proposed by Raykar et al (2010) (denoted as EM-R), the scalar Simultaneous Truth and Performance Level Estimation (denoted as sSTAPLE) proposed by Warfield et al (2008), and also the mean and median voting approaches.…”
Section: Capnobase Rr Datasetmentioning
confidence: 99%
“…In addition, BCLAs was compared to BCLA, 'smart fusion' (i.e. the benchmarking algorithm proposed by authors of the dataset (Karlen et al 2013)), the bestperforming (lowest-MAE) single algorithm (denoted 'Best'), the EM algorithm proposed by Raykar et al (2010) (denoted as EM-R), the scalar Simultaneous Truth and Performance Level Estimation (denoted as sSTAPLE) proposed by Warfield et al (2008), and also the mean and median voting approaches.…”
Section: Capnobase Rr Datasetmentioning
confidence: 99%
“…The growth of crowdwork applications has catalyzed interest in two related but distinct research questions concerning human evaluations: inferring true labels from multiple annotations [1]- [7] and estimating evaluator expertise [8]- [12].…”
Section: Related Workmentioning
confidence: 99%
“…There have been attempts at unifying the two research directions by modeling evaluators and items [9,15,16,17]. [9] associates an expertise level with individual evaluators and a difficulty level with each of the items; the probability that an evaluator rates an item correctly is then modeled as a function of these two parameters.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations