2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00788
|View full text |Cite
|
Sign up to set email alerts
|

RGBD Based Dimensional Decomposition Residual Network for 3D Semantic Scene Completion

Abstract: RGB images differentiate from depth as they carry more details about the color and texture information, which can be utilized as a vital complement to depth for boosting the performance of 3D semantic scene completion (SSC). SSC is composed of 3D shape completion (SC) and semantic scene labeling while most of the existing approaches use depth as the sole input which causes the performance bottleneck. Moreover, the state-of-the-art methods employ 3D CNNs which have cumbersome networks and tremendous parameters.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
116
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 80 publications
(116 citation statements)
references
References 49 publications
0
116
0
Order By: Relevance
“…We validate our approach on both the synthetic SUNCG-RGBD dataset (Liu et al 2018) and the real NYUv2 dataset (Silberman et al 2012), and the results show that our method achieves a gain of 2.5% on the SUNCG-RGBD dataset and a gain of 2.6% on the NYUv2 dataset against the state-of-the-art method (Li et al 2019b). Our analytical study also observes that directly boosting the AMFNet from the 2D segmentation ground truth of the input RGB-D images (also see Figure 1(f)) can achieve a gain of 3.8% on the synthetic SUNCG-RGBD dataset and a gain of 4.3% on the real NYUv2 dataset against the state-of-the-art method (Li et al 2019b).…”
Section: Introductionmentioning
confidence: 88%
See 3 more Smart Citations
“…We validate our approach on both the synthetic SUNCG-RGBD dataset (Liu et al 2018) and the real NYUv2 dataset (Silberman et al 2012), and the results show that our method achieves a gain of 2.5% on the SUNCG-RGBD dataset and a gain of 2.6% on the NYUv2 dataset against the state-of-the-art method (Li et al 2019b). Our analytical study also observes that directly boosting the AMFNet from the 2D segmentation ground truth of the input RGB-D images (also see Figure 1(f)) can achieve a gain of 3.8% on the synthetic SUNCG-RGBD dataset and a gain of 4.3% on the real NYUv2 dataset against the state-of-the-art method (Li et al 2019b).…”
Section: Introductionmentioning
confidence: 88%
“…Given a single depth image or RGB-D images of a 3D scene, many papers (Gupta, Arbelaez, and Malik 2013;Ren, Bo, and Fox 2012;Firman et al 2016) have been proposed to complete or segment the 3D scene with neural networks. More recently, a set of methods (Song et al 2017;Liu et al 2018;Guo and Tong 2018;Zhang et al 2018;Li et al 2019b) have been developed to automatically predict the semantic labels, together with completing the 3D geometry, of the objects in a 3D scene from a single view of the scene using convolutional neural networks.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…To enrich the input information and boost the accuracy of SSC, TS3D [12] and DDR-SSC [17] proposed to add a RGB branch in addition to the voxel branch, which introduce extra network or parameters, and are less accurate than our method.…”
Section: A Semantic Scene Completionmentioning
confidence: 99%