2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015
DOI: 10.1109/cvpr.2015.7298824
|View full text |Cite
|
Sign up to set email alerts
|

Best of both worlds: Human-machine collaboration for object annotation

Abstract: The long-standing goal of localizing every object in an image remains elusive. Manually annotating objects is quite expensive despite crowd engineering innovations. Current state-of-the-art automatic object detectors can accurately detect at most a few objects per image. This paper brings together the latest advancements in object detection and in crowd engineering into a principled framework for accurately and efficiently localizing objects in images. The input to the system is an image to annotate and a set … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
129
0
1

Year Published

2016
2016
2020
2020

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 193 publications
(132 citation statements)
references
References 54 publications
(116 reference statements)
2
129
0
1
Order By: Relevance
“…many ML-methods perform very badly on extrapolation problems which would be very easy for humans [32, p. 4] -and major assumptions of normative models cannot be applied in reality, a conjoint approach of human and machine input could be key to enhanced decision quality. Therefore, the answer is to put humans in the loop [40]. However, using normative models to integrate human decision making in centrals parts of machine learning could lead to faulty predictions since the nature of actual decision making is of bounded rationality [5].…”
Section: Open Problemsmentioning
confidence: 99%
See 1 more Smart Citation
“…many ML-methods perform very badly on extrapolation problems which would be very easy for humans [32, p. 4] -and major assumptions of normative models cannot be applied in reality, a conjoint approach of human and machine input could be key to enhanced decision quality. Therefore, the answer is to put humans in the loop [40]. However, using normative models to integrate human decision making in centrals parts of machine learning could lead to faulty predictions since the nature of actual decision making is of bounded rationality [5].…”
Section: Open Problemsmentioning
confidence: 99%
“…One important problem which we have to face in future research is which questions to pose to humans and how to ask those questions [40]. At this point, human machineinteraction could provide useful insights and offer guidelines for the design of interfaces and visualisations.…”
Section: Open Problemsmentioning
confidence: 99%
“…tools for efficient video annotation [53] and object labelling [34]. Methods that intelligently design the query space [39,32,30] also share the spirit of reducing annotation effort. Other works have looked into active learning schemes that query for multiple types of annotator feedback [50,4,43].…”
Section: Related Workmentioning
confidence: 99%
“…integrate computer vision and crowdsourcing is to understand the performance of computer vision algorithms and adaptively use crowd work [158,215]. In chapter 5, we described a method to use an ML-based supervised workflow controller to assess the difficulty of each object detection task, which allowed us to adaptively allocate work to different crowd workflows to reduce human cost.…”
Section: Combining Of Crowdsourcing and Computer Vision The Key To Ementioning
confidence: 99%
“…For example, Zhang et al used a supervised machine learning algorithms to detect computer vision failure in semantic image segmentation and vanishing point detection [215]. Russakovsky et al introduced a method to combine a variety of crowdsourcing tasks with object detection algorithms using Markov decision process that automatically balances human cost and object detection accuracy [158].…”
Section: Combining Of Crowdsourcing and Computer Vision The Key To Ementioning
confidence: 99%