2020
DOI: 10.1007/978-3-030-58529-7_34
|View full text |Cite
|
Sign up to set email alerts
|

Feature-Metric Loss for Self-supervised Learning of Depth and Egomotion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
141
0
1

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 184 publications
(142 citation statements)
references
References 47 publications
0
141
0
1
Order By: Relevance
“…multi-scale prediction to prevent the training target from being trapped in the local minimum with gradient locality by bilinear sampling. Recent approaches add loss [33], networks such as an optical flow network for motion information supplementation [34,35], and a feature-metric network for semantic information addition [36] and reduce the performance difference between monocular and stereo-based depth estimation. However, this unsupervised learned depth is not guaranteed by a metric measure.…”
Section: Self-supervised Trainingmentioning
confidence: 99%
See 2 more Smart Citations
“…multi-scale prediction to prevent the training target from being trapped in the local minimum with gradient locality by bilinear sampling. Recent approaches add loss [33], networks such as an optical flow network for motion information supplementation [34,35], and a feature-metric network for semantic information addition [36] and reduce the performance difference between monocular and stereo-based depth estimation. However, this unsupervised learned depth is not guaranteed by a metric measure.…”
Section: Self-supervised Trainingmentioning
confidence: 99%
“…Since the depth discontinuity depends on the gradients δI t of the image, the edgeaware term is used together as in previous studies [17,36,37] to limit the high depth gradient δD t for the texture-less region.…”
Section: Depth Smoothness Lossmentioning
confidence: 99%
See 1 more Smart Citation
“…With advances in deep networks, Eigen et al [2] showed that monocular depth estimation with deep neural network yields significant gains in accuracy and speed compared to the traditional attempts with handcrafted features. Starting with the work of Eigen et al [2], many learning-based methods [1,3,4,5,6,7,8,9,10,11] have been studied and showed high accuracy in overall depth evaluation metrics. However, they still produce noisy and temporally flickering depth maps because they perform depth estimation on each frame independently or do not use temporal information correctly.…”
Section: Introductionmentioning
confidence: 99%
“…Several video-based depth estimation methods have been studied, and the core idea behind these methods is to utilize temporal information. Temporal consistency can be explicitly constrained [1,7,8,9,12,13] or implicitly constrained using the recurrent neural networks [7,10,11]. Eom et al [7] proposed the recurrent model with a flow-guided memory unit for consistent depth.…”
Section: Introductionmentioning
confidence: 99%