2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00032
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning of Volume-Guided Progressive View Inpainting for 3D Point Scene Completion From a Single Depth Image

Abstract: We present a deep reinforcement learning method of progressive view inpainting for 3D point scene completion under volume guidance, achieving high-quality scene reconstruction from only a single depth image with severe occlusion. Our approach is end-to-end, consisting of three modules: 3D scene volume reconstruction, 2D depth map inpainting, and multi-view selection for completion. Given a single depth image, our method first goes through the 3D volume branch to obtain a volumetric scene reconstruction as a gu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
26
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 48 publications
(26 citation statements)
references
References 56 publications
0
26
0
Order By: Relevance
“…Another approach uses a 3D encoder–predictor convoluted neural network that combines a 3D deep learning architecture with a 3D shape synthesis technique [ 51 ]. Whereas, other approaches complete scenes by inferring points from depth maps [ 12 , 53 ]. ScanComplete [ 8 ]—a large-scale scene completion method from 3D scans—uses fully convoluted neural networks that can be trained on smaller sets and applied to larger scenes, which allows for efficient processing.…”
Section: Literature Reviewmentioning
confidence: 99%
See 1 more Smart Citation
“…Another approach uses a 3D encoder–predictor convoluted neural network that combines a 3D deep learning architecture with a 3D shape synthesis technique [ 51 ]. Whereas, other approaches complete scenes by inferring points from depth maps [ 12 , 53 ]. ScanComplete [ 8 ]—a large-scale scene completion method from 3D scans—uses fully convoluted neural networks that can be trained on smaller sets and applied to larger scenes, which allows for efficient processing.…”
Section: Literature Reviewmentioning
confidence: 99%
“…However, to apply these methods, additional segmentation is required to extract individual objects from a larger scene. Recently, methods using deep learning have been developed [8,9,12,[49][50][51][52][53], and some extend beyond single objects to perform semantic scene segmentation [8,9,12,53]. One approach focused on point cloud upsampling [50], creating denser and more uniform sets of points from a sparse input.…”
Section: Data-driven 3d Scene Completionmentioning
confidence: 99%
“…Song et al (2017) used an end-to-end network SSCNet for scene completion and Guo and Tong (2018) a view-volume CNN that extracts geometric features from 2D depth images. Zhang and Funkhouser (2018) presented an end-to-end architecture for depth inpainting, and Han et al (2019) used multi-view depth completion to predict point cloud representations. A 3D recurrent network has been used to integrate information from only a few input views (Choy et al, 2016), and Xu et al (2016) used spatial and temporal structure of sequential observations to predict a view sequence.…”
Section: Introductionmentioning
confidence: 99%
“…Understanding and reconstructing a 3D scene from partial observations is a very important technique which has received increasing research attention in recent years due to its commercial potential in a large variety of robotics and vision tasks such as robotic navigation (Gupta, Arbelaez, and Malik 2013), autonomous driving (Laugier et al 2011), and scene reconstruction (Hays and Efros 2007;Han et al 2019). Given a single depth image or RGB-D images of a 3D scene, many papers (Gupta, Arbelaez, and Malik 2013;Ren, Bo, and Fox 2012;Firman et al 2016) have been proposed to complete or segment the 3D scene with neural networks.…”
Section: Introductionmentioning
confidence: 99%