Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence 2023
DOI: 10.24963/ijcai.2023/155
|View full text |Cite
|
Sign up to set email alerts
|

Depth-Relative Self Attention for Monocular Depth Estimation

Abstract: Monocular depth estimation is very challenging because clues to the exact depth are incomplete in a single RGB image. To overcome the limitation, deep neural networks rely on various visual hints such as size, shade, and texture extracted from RGB information. However, we observe that if such hints are overly exploited, the network can be biased on RGB information without considering the comprehensive view. We propose a novel depth estimation model named RElative Depth Transformer (RED-T) that uses relative de… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…Consequently, current MDE models usually overfit to specific datasets, making them less effective in generalizing to other datasets. In RDE, the consistency of depth predictions is limited to within image frames, and the scale factor remains unknown [5]. This unique characteristic enables training methods on various scenes and datasets, including 3D movies, promoting model generalizability across different domains.…”
Section: General Information About Monocular Depth Estimationmentioning
confidence: 99%
“…Consequently, current MDE models usually overfit to specific datasets, making them less effective in generalizing to other datasets. In RDE, the consistency of depth predictions is limited to within image frames, and the scale factor remains unknown [5]. This unique characteristic enables training methods on various scenes and datasets, including 3D movies, promoting model generalizability across different domains.…”
Section: General Information About Monocular Depth Estimationmentioning
confidence: 99%