2019
DOI: 10.1609/aaai.v33i01.33018001
|View full text |Cite
|
Sign up to set email alerts
|

Depth Prediction without the Sensors: Leveraging Structure for Unsupervised Learning from Monocular Videos

Abstract: Learning to predict scene depth from RGB inputs is a challenging task both for indoor and outdoor robot navigation. In this work we address unsupervised learning of scene depth and robot ego-motion where supervision is provided by monocular videos, as cameras are the cheapest, least restrictive and most ubiquitous sensor for robotics. Previous work in unsupervised image-to-depth learning has established strong baselines in the domain. We propose a novel approach which produces higher quality results, is able t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
396
2

Year Published

2019
2019
2020
2020

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 452 publications
(411 citation statements)
references
References 19 publications
2
396
2
Order By: Relevance
“…The CNNs is trained by a photometric reconstruction loss, which was obtained by warping nearby views to the target using the computed depth and pose. In addition, [21] addresses unsupervised learning of scene depth, robot ego-motion and object motions where the supervision is provided by geometric structure of monocular videos as input. Furthermore, many recent efforts [11], [22] explore the geometric relationships between depth, camera motion, and flow for unsupervised learning of depth and flow estimation models.…”
Section: Unsupervised Learning Methodsmentioning
confidence: 99%
“…The CNNs is trained by a photometric reconstruction loss, which was obtained by warping nearby views to the target using the computed depth and pose. In addition, [21] addresses unsupervised learning of scene depth, robot ego-motion and object motions where the supervision is provided by geometric structure of monocular videos as input. Furthermore, many recent efforts [11], [22] explore the geometric relationships between depth, camera motion, and flow for unsupervised learning of depth and flow estimation models.…”
Section: Unsupervised Learning Methodsmentioning
confidence: 99%
“…Vijayanarasimhan et al [41] additionally learn rigid motion parameters of multiple objects. Subsequent methods further proposed improvements based on various techniques, such as an ICP alignment loss [30], supervision from SfM algorithms [26], optical flow [48,52], edges [47], modeling multiple rigid motions informed by instance segmentation [3], motion segmentation [36], minimum projection loss [16], etc.…”
Section: Related Workmentioning
confidence: 99%
“…Frames that are similar to the test scenes are removed from the training set. We compare the performance of the proposed framework with the baseline, as well as recent stateof-the-art works in the same setting [3,30,42,47,48,51,52].…”
Section: Depth Estimationmentioning
confidence: 99%
See 2 more Smart Citations