2019 International Conference on Robotics and Automation (ICRA) 2019
DOI: 10.1109/icra.2019.8793515
|View full text |Cite
|
Sign up to set email alerts
|

One-Shot Learning of Multi-Step Tasks from Observation via Activity Localization in Auxiliary Video

Abstract: Due to burdensome data requirements, learning from demonstration often falls short of its promise to allow users to quickly and naturally program robots. Demonstrations are inherently ambiguous and incomplete, making correct generalization to unseen situations difficult without a large number of demonstrations in varying conditions. By contrast, humans are often able to learn complex tasks from a single demonstration (typically observations without action labels) by leveraging context learned over a lifetime. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
18
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 25 publications
(18 citation statements)
references
References 20 publications
0
18
0
Order By: Relevance
“…Simulation has also been leveraged as supervision to learn such representations [32] or to produce human data with domain randomization [3]. Finally, meta-learning [54] and subtask discovery [41,20] have also been explored as techniques for acquiring robot rewards or demos from human videos. In contrast to the majority of these works, which usually study a small set of human videos in a similar domain as the robot, we explicitly focus on leveraging "in-the-wild" human videos, specifically large and diverse sets of crowd-sourced videos from the real world from an existing dataset, which contains many different individuals, viewpoints, backgrounds, objects, and tasks.…”
Section: B Robotic Learning From Human Videosmentioning
confidence: 99%
“…Simulation has also been leveraged as supervision to learn such representations [32] or to produce human data with domain randomization [3]. Finally, meta-learning [54] and subtask discovery [41,20] have also been explored as techniques for acquiring robot rewards or demos from human videos. In contrast to the majority of these works, which usually study a small set of human videos in a similar domain as the robot, we explicitly focus on leveraging "in-the-wild" human videos, specifically large and diverse sets of crowd-sourced videos from the real world from an existing dataset, which contains many different individuals, viewpoints, backgrounds, objects, and tasks.…”
Section: B Robotic Learning From Human Videosmentioning
confidence: 99%
“…While, third-person imitation learning uses date from other agents or viewpoints [27,35]. Recent methods for one-shot imitation learning [8,11,13,40,41,42] can translate a single demonstration to an executable pol- icy. The most similar to ours is NTP [41] that also learns long-horizon tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Prior work in LfD has tackled the challenging problem of extracting task plans from a single end-user demonstration [10], [11], [12], [13], [14], [15], [16], [17], [18]. These approaches present intuitive ways for end-users to program complex robot behaviors using kinesthetic teaching [10], virtual reality [11], GUI programming [12], or direct demonstration [18].…”
Section: Related Workmentioning
confidence: 99%