Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence 2021
DOI: 10.24963/ijcai.2021/599
|View full text |Cite
|
Sign up to set email alerts
|

Understanding the Relationship between Interactions and Outcomes in Human-in-the-Loop Machine Learning

Abstract: Human-in-the-loop Machine Learning (HIL-ML) is a widely adopted paradigm for instilling human knowledge in autonomous agents. Many design choices influence the efficiency and effectiveness of such interactive learning processes, particularly the interaction type through which the human teacher may provide feedback. While different interaction types (demonstrations, preferences, etc.) have been proposed and evaluated in the HIL-ML literature, there has been little discussion of how these compare or how they sho… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 19 publications
(15 citation statements)
references
References 50 publications
0
14
0
1
Order By: Relevance
“…1 Supervised learning covers a broad range of model classes, ranging from vision transformers, large language models, and impactful application areas, like medical diagnostics [69] and criminal justice [99]. We consider other objectives, which may include reinforcement learning [23,30] or unsupervised learning [24,55], to be out-of-scope for this paper. We focus on methods more commonly deployed in practice, and omit the Bayesian analog for our pipeline, where experts can express preferences over the distribution of functions [56,97].…”
Section: Feedback-update Taxonomymentioning
confidence: 99%
See 1 more Smart Citation
“…1 Supervised learning covers a broad range of model classes, ranging from vision transformers, large language models, and impactful application areas, like medical diagnostics [69] and criminal justice [99]. We consider other objectives, which may include reinforcement learning [23,30] or unsupervised learning [24,55], to be out-of-scope for this paper. We focus on methods more commonly deployed in practice, and omit the Bayesian analog for our pipeline, where experts can express preferences over the distribution of functions [56,97].…”
Section: Feedback-update Taxonomymentioning
confidence: 99%
“…Before deploying a machine learning (ML) model in high-stakes use cases, practitioners, who are responsible for developing and maintaining models, may solicit and incorporate feedback from experts [4,30,44]. Prior work has largely focused on incorporating feedback of technical experts (e.g., from ML engineers, data scientists) into models [2,90,92,107,114,116,124].…”
Section: Introductionmentioning
confidence: 99%
“…Human Feedback: This describes the feedback that the teacher will provide to the machine. [Cui et al, 2021] defined four broad types of feedback i.e. showing, categorizing, sorting and evaluating.…”
Section: Teaching Interfacementioning
confidence: 99%
“…Human-in-the-loop (HITL) ML: [Cui et al, 2021] surveyed HITL ML and how design choices affect interactive learning. In their framework, they define four types of interaction: Showing, Categorizing, Sorting and Evaluating.…”
Section: Ablation Studiesmentioning
confidence: 99%
“…Finally, this paper has explored one method of interaction to enable a human teacher to provide corrections to the robot. However, in human-in-the-loop learning problems, the ideal interaction type is dependent on the teacher's role in the learning system, and the context in which the robot is used (Cui et al, 2021). For example, the teacher may not have time to correct every step of the robot's action, or may instead prefer to provide corrections only after the robot has tried and failed to complete a task.…”
Section: Open Questionsmentioning
confidence: 99%