2017
DOI: 10.48550/arxiv.1703.11000
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning Visual Servoing with Deep Features and Fitted Q-Iteration

Alex X. Lee,
Sergey Levine,
Pieter Abbeel

Abstract: Visual servoing involves choosing actions that move a robot in response to observations from a camera, in order to reach a goal configuration in the world. Standard visual servoing approaches typically rely on manually designed features and analytical dynamics models, which limits their generalization capability and often requires extensive application-specific feature and model engineering.In this work, we study how learned visual features, learned predictive dynamics models, and reinforcement learning can be… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(13 citation statements)
references
References 19 publications
0
13
0
Order By: Relevance
“…The first category of tasks is ones where the goal location is known, and limited exploration is necessary. This could be in the form of a simply wandering around without colliding [15,34], following an object [23], getting to a goal coordinate [1,17]: using sequence of images along the path [5,22], or language-based instructions [2]. Sometimes, the goal is specified as an image but experience from the environment is available in the form of demonstrations [13,35], or in the form of reward-based training [25,48], which again limits the role of exploration.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The first category of tasks is ones where the goal location is known, and limited exploration is necessary. This could be in the form of a simply wandering around without colliding [15,34], following an object [23], getting to a goal coordinate [1,17]: using sequence of images along the path [5,22], or language-based instructions [2]. Sometimes, the goal is specified as an image but experience from the environment is available in the form of demonstrations [13,35], or in the form of reward-based training [25,48], which again limits the role of exploration.…”
Section: Related Workmentioning
confidence: 99%
“…Depending on the problem being considered, different representations have been investigated. For short-range locomotion tasks, purely reactive policies [3,15,23,34] suffice. For more complex problems such as target-driven navigation in a novel environment, such purely reactive strategies do not work well [48], and memory-based policies have been investigated.…”
Section: Related Workmentioning
confidence: 99%
“…[10], instead, manages to train from single view image streams a neural network able to predict the probability of successful grasps, learning thus a hand-eye coordination for grasping. Interesting related works on visual DRL for robotics are also [11], [12], [13], [14]. Data efficient DRL for DPG-based dexterous manipulation has been further explored in [15], which mainly focuses on stacking Lego blocks.…”
Section: State Of the Artmentioning
confidence: 99%
“…Visual servoing: Finally, there have been multiple approaches to visual servoing over the years [1], [17], [18], including some newer methods that use deep learned features and reinforcement learning [19]. While these methods depend on an external system for data association or on pre-specified features, our system is trained end-to-end and can control directly from raw depth data.…”
Section: Related Workmentioning
confidence: 99%