2021
DOI: 10.1109/lra.2021.3062602
|View full text |Cite
|
Sign up to set email alerts
|

An Adaptive Framework for Learning Unsupervised Depth Completion

Abstract: We present a method to infer a dense depth map from a color image and associated sparse depth measurements. Our main contribution lies in the design of an annealing process for determining co-visibility (occlusions, disocclusions) and the degree of regularization to impose on the model. We show that regularization and co-visibility are related via the fitness (residual) of model to data and both can be unified into a single framework to improve the learning process. Our method is an adaptive weighting scheme t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 29 publications
(8 citation statements)
references
References 32 publications
0
8
0
Order By: Relevance
“…Both stereo [26], [36] and monocular [21], [31], [32], [33] paradigms learn dense depth from an image and sparse depth measurements by minimizing the photometric error between the input image and its reconstruction from other views along with the difference between prediction and sparse depth input (sparse depth reconstruction). [21] used Perspective-n-Point [19] and RANSAC [9] to align consecutive frames and [34], [32] proposed an adaptive weighting framework. [36] also used synthetic data but require image, sparse and ground-truth depth.…”
Section: Related Workmentioning
confidence: 99%
“…Both stereo [26], [36] and monocular [21], [31], [32], [33] paradigms learn dense depth from an image and sparse depth measurements by minimizing the photometric error between the input image and its reconstruction from other views along with the difference between prediction and sparse depth input (sparse depth reconstruction). [21] used Perspective-n-Point [19] and RANSAC [9] to align consecutive frames and [34], [32] proposed an adaptive weighting framework. [36] also used synthetic data but require image, sparse and ground-truth depth.…”
Section: Related Workmentioning
confidence: 99%
“…Selfsupervised or unsupervised depth completion assumes that the ground-truth dense depth is not available. To address this task, a main solution is minimizing the depth loss between the predicted depth and sparse map, and the photometric loss between the RBG image and its warped image [11], [12], [29]- [31]. For example, Ma et al [29] utilized Perspective-n-Point (PnP) [32] and Random Sample Consensus (RANSAC) [33] to estimate the relative pose between two adjacent frames and warp image.…”
Section: Related Workmentioning
confidence: 99%
“…Unsupervised/Self-supervised depth completion assumes stereo images or monocular videos to be available during training. Both stereo [38,49] and monocular [25,44,45,46] training paradigms leverage sparse depth reconstruction and photometric reprojection error as a training signal by minimizing photometric discrepancies between the input image and its reconstruction from other views. [25] used Perspective-n-Point [22] and RANSAC [11] to align consecutive video frames.…”
Section: Related Work and Contributionsmentioning
confidence: 99%
“…Motivation. Unsupervised methods [25,44,45,46] use the photometric reprojection error perl as a training signal. The input image I t is reconstructed from temporally adjacent frames…”
Section: Kbnet Architecturementioning
confidence: 99%
See 1 more Smart Citation