2020
DOI: 10.48550/arxiv.2002.12319
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Semantically-Guided Representation Learning for Self-Supervised Monocular Depth

Abstract: Self-supervised learning is showing great promise for monocular depth estimation, using geometry as the only source of supervision. Depth networks are indeed capable of learning representations that relate visual appearance to 3D properties by implicitly leveraging category-level patterns. In this work we investigate how to leverage more directly this semantic structure to guide geometric representation learning, while remaining in the self-supervised regime. Instead of using semantic labels and proxy losses i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
37
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 22 publications
(37 citation statements)
references
References 33 publications
0
37
0
Order By: Relevance
“…It demonstrates that when the fine-tuning dataset is similar to the test dataset, the finetuning method performs better than the data transformation method. The best domain adaptation method [138] has superior performance to the best unsupervised method [119]. Regarding the best semi-supervised method [119] and the best domain adaptation method [138], [119] outperforms [138].…”
Section: A Accuracymentioning
confidence: 95%
See 2 more Smart Citations
“…It demonstrates that when the fine-tuning dataset is similar to the test dataset, the finetuning method performs better than the data transformation method. The best domain adaptation method [138] has superior performance to the best unsupervised method [119]. Regarding the best semi-supervised method [119] and the best domain adaptation method [138], [119] outperforms [138].…”
Section: A Accuracymentioning
confidence: 95%
“…The best domain adaptation method [138] has superior performance to the best unsupervised method [119]. Regarding the best semi-supervised method [119] and the best domain adaptation method [138], [119] outperforms [138]. In particular, their accuracy metrics are almost equal, while the error metrics especially Sq Rel, RMSE and RMSE log are highly variable.…”
Section: A Accuracymentioning
confidence: 96%
See 1 more Smart Citation
“…Our weakly-supervised method achieves competitive subjective results with fully-supervised ones. [49], [50], [51], [47], [46], [54], semi-supervised [55], [56], [54], [9], [26], [10] and fully-supervised ones [1], [19], [2], [6], [3], [21], [4], [5], [22], [23], [14], [44], [45], [24] on two benchmark datasets, i.e., KITTI [42] and NYU Depth-v2 [52] datasets. Since the fully-supervised methods exploit the groundtruth HR depth maps for training, in theory it should work better than other learning manners.…”
Section: Performance Comparisonmentioning
confidence: 99%
“…Guizilini et al [13] learned to compress and decompress detail-preserving representations by symmetrical packing and unpacking blocks. Other published methods were based upon edge and normal [48,49], Competitive Collaboration [36], semantic segmentation [14,22] and feature representations learning [41]. A state-of-theart framework was Monodepth2 proposed by Godard et al [11], which introduced a minimum re-projection loss to deal with occlusions and auto-masking scheme removing invalid pixels robustly.…”
Section: Self-supervised Monocular Depth Estimationmentioning
confidence: 99%