2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00438
|View full text |Cite
|
Sign up to set email alerts
|

Pay Attention! - Robustifying a Deep Visuomotor Policy Through Task-Focused Visual Attention

Abstract: Several recent studies have demonstrated the promise of deep visuomotor policies for robot manipulator control. Despite impressive progress, these systems are known to be vulnerable to physical disturbances, such as accidental or adversarial bumps that make them drop the manipulated object. They also tend to be distracted by visual disturbances such as objects moving in the robot's field of view, even if the disturbance does not physically prevent the execution of the task. In this paper, we propose an approac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 29 publications
(25 citation statements)
references
References 25 publications
0
25
0
Order By: Relevance
“…As mentioned above, the robustness and generalization of these end-to-end learning-based methods are limited by the range of experience in the training phase. Abolghasemi et al applied a task-focused visual attention mechanism in the end-to-end IL framework to enhance the robustness of visuomotor manipulation skills [22]. James et al trained a visuomotor skill for pick-and-place task in simulation with domain randomization and transferred the visuomotor skill to the real world without any fine-tuning [3].…”
Section: Related Workmentioning
confidence: 99%
“…As mentioned above, the robustness and generalization of these end-to-end learning-based methods are limited by the range of experience in the training phase. Abolghasemi et al applied a task-focused visual attention mechanism in the end-to-end IL framework to enhance the robustness of visuomotor manipulation skills [22]. James et al trained a visuomotor skill for pick-and-place task in simulation with domain randomization and transferred the visuomotor skill to the real world without any fine-tuning [3].…”
Section: Related Workmentioning
confidence: 99%
“…It was also used for motion switching in multi-step task such as cloth-folding [45]. Abolghasemi et al improved the robustness of this multi-task approach via task-focused visual attention [15]. Hausman et al also combined an embedding task space with RL to learn transferable robot skills [46].…”
Section: Related Workmentioning
confidence: 99%
“… res is the deviation of the observation o and the restored o , , which is expressed with the square of Euclidean Norm and  VAE is the total loss of VAE. Inspired by the work of Abolghasemi [15], another loss term that is the deviation of the reconstructed images of encoded vector is added to enhance coding performance.…”
Section: Perception Networkmentioning
confidence: 99%
See 2 more Smart Citations