Robotics: Science and Systems XIV 2018
DOI: 10.15607/rss.2018.xiv.009
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement and Imitation Learning for Diverse Visuomotor Skills

Abstract: We propose a model-free deep reinforcement learning method that leverages a small amount of demonstration data to assist a reinforcement learning agent. We apply this approach to robotic manipulation tasks and train end-to-end visuomotor policies that map directly from RGB camera inputs to joint velocities. We demonstrate that our approach can solve a wide variety of visuomotor tasks, for which engineering a scripted controller would be laborious. In experiments, our reinforcement and imitation agent achieves … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
118
0
1

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 201 publications
(120 citation statements)
references
References 49 publications
1
118
0
1
Order By: Relevance
“…In these systems, an artificial agent must learn to produce action outcomes in response to information from the environment, including rewards. Several such artificial systems have used convolutional architectures on the front end in order to process visual information about the world (Figure 3) [72,73,74]. It would be interesting to compare the representations learned in the context of these models to those trained by other mechanisms, as well as to data.…”
Section: Alternative Training Proceduresmentioning
confidence: 99%
“…In these systems, an artificial agent must learn to produce action outcomes in response to information from the environment, including rewards. Several such artificial systems have used convolutional architectures on the front end in order to process visual information about the world (Figure 3) [72,73,74]. It would be interesting to compare the representations learned in the context of these models to those trained by other mechanisms, as well as to data.…”
Section: Alternative Training Proceduresmentioning
confidence: 99%
“…Reinforcement learning approaches have recently been proposed to address variations in geometry and configuration for manipulation. [34,58] trained neural network policies using RGB images and proprioceptive feedback. Their approach works well in a wide range of tasks, but the large object clearances compared to automation tasks may explain the sufficiency of RGB data.…”
Section: A Contact-rich Manipulationmentioning
confidence: 99%
“…While the utility of multimodal data has frequently been shown in robotics [7,45,48,54], the proposed manipulation strategies are often task-specific. While learning-based methods do not require manual task specification, the majority of learned manipulation policies close the control loop around a single modality, often vision [15,21,34,58].…”
Section: Introductionmentioning
confidence: 99%
“…Reinforcement learning approaches have recently been proposed to address variations in geometry and configuration for manipulation. [41,72] train neural network policies using RGB images and proprioceptive feedback. Their approach works well in a wide range of tasks, but the large object clearances compared to manufacturing tasks may explain the sufficiency of RGB data.…”
Section: A Contact-rich Manipulationmentioning
confidence: 99%
“…This makes many of these methods task-specific. On the other hand, most learning-based methods do not require manual task specification, yet the majority of learned manipulation policies close the control loop around a single modality, often RGB images [15,22,41,72].…”
Section: Introductionmentioning
confidence: 99%