2021
DOI: 10.1109/lra.2021.3095269
|View full text |Cite
|
Sign up to set email alerts
|

Uncertainty-Aware Self-Supervised Learning of Spatial Perception Tasks

Abstract: We propose a general self-supervised learning approach for spatial perception tasks, such as estimating the pose of an object relative to the robot, from onboard sensor readings. The model is learned from training episodes, by relying on: a continuous state estimate, possibly inaccurate and affected by odometry drift; and a detector, that sporadically provides supervision about the target pose. We demonstrate the general approach in three different concrete scenarios: a simulated robot arm that visually estima… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
15
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3

Relationship

3
4

Authors

Journals

citations
Cited by 15 publications
(15 citation statements)
references
References 23 publications
0
15
0
Order By: Relevance
“…Zeng et al [10] perform multi-view object segmentation to generate training data for object pose estimation. Nava et al [22] exploit noisy state estimates to assist self-supervised learning of spatial perception models.…”
Section: B Multi-view Object-based Perceptionmentioning
confidence: 99%
“…Zeng et al [10] perform multi-view object segmentation to generate training data for object pose estimation. Nava et al [22] exploit noisy state estimates to assist self-supervised learning of spatial perception models.…”
Section: B Multi-view Object-based Perceptionmentioning
confidence: 99%
“…the drone. Perception approaches for robotic arms focus on identifying and localizing an object to be grasped [1], [21], [22], [23]. For example, Pinto et al [1] predict the probability of success of a grasp attempt, considering the optimal angle for approaching the object.…”
Section: Related Workmentioning
confidence: 99%
“…With a similar camera configuration, Tobin et al [21] learn to estimate the location of known geometrical objects (e.g., pyramids, cones, cylinders, and cubes); Zeng et al [22] estimate the full 3D pose of objects from the feed of multiple fixed-inspace cameras. In contrast, Nava et al [23] learn the 3D pose of objects using only an uncalibrated monocular camera attached to the end-effector.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations