2020
DOI: 10.1007/978-3-030-58589-1_31
|View full text |Cite
|
Sign up to set email alerts
|

Monocular Differentiable Rendering for Self-supervised 3D Object Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
21
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 37 publications
(22 citation statements)
references
References 28 publications
0
21
0
Order By: Relevance
“…The current research methods for 3D target detection can be summarized as LIDAR [3][4] based detection methods, and depth image [5][6] based detection methods. The depth image detection methods are classified into monocular [7][8], binocular [9], and multiocular 3D target detection by designing 2D image feature extractors to capture pixels. In addition, some studies combined multiple heterogeneous feature map with a fusion mode view study approach [10] [11], which proved the superior performance of the fusion module by enhancing point-by-point features with semantic image features [12].…”
Section: Related Workmentioning
confidence: 99%
“…The current research methods for 3D target detection can be summarized as LIDAR [3][4] based detection methods, and depth image [5][6] based detection methods. The depth image detection methods are classified into monocular [7][8], binocular [9], and multiocular 3D target detection by designing 2D image feature extractors to capture pixels. In addition, some studies combined multiple heterogeneous feature map with a fusion mode view study approach [10] [11], which proved the superior performance of the fusion module by enhancing point-by-point features with semantic image features [12].…”
Section: Related Workmentioning
confidence: 99%
“…[22] for a comprehensive overview. This spawned several applications like self-supervision for monocular 3D object detection [5]. In [16], a differentiable renderer is employed -and even learned -to predict geometric correspondence fields to refine pose estimates of 3D objects.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, many works [2,20,24] concentrate on vehicle 3D texture recovery under real environments. Due to the lack of ground truth 3D data of real scene, they mainly pay attention to reconstruct 3D models from 2D data utilizing unsupervised or self-supervised learning.…”
Section: Monocular Vehicle Reconstructionmentioning
confidence: 99%
“…Monocular visual scene understanding, which mainly focuses on high-level understanding of single image content, is a fundamental technology for many automatic applications, especially in the field of autonomous driving. Using only a single-view driving image, available vehicle parsing studies have covered popular topics starting from 2D vehicle detection [3,34,32,14,31,46], then 6D vehicle pose recovery [59,55,35,26,52,12,13,4,33], and finally vehicle shape reconstruction [10,28,50,20,2,24,15,29,36,61]. However, much less efforts are devoted to vehicle texture estimation.…”
Section: Introductionmentioning
confidence: 99%