2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.269
|View full text |Cite
|
Sign up to set email alerts
|

Synthesizing 3D Shapes via Modeling Multi-view Depth Maps and Silhouettes with Deep Generative Networks

Abstract: Our method generates multi-view depth maps and silhouettes, and uses a rendering function to obtain the 3D shapes. Right: We can also extend our framework to reconstruct 3D shapes from single/multi-view depth maps or silhouettes.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
152
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 182 publications
(152 citation statements)
references
References 20 publications
0
152
0
Order By: Relevance
“…Most of the missclassification comes from the similar category pairs like (table, desk ) and (night stand, dresser ). We also perform better than all the appearance based methods, except Wang et al [27] which performs a specialized view based ModelNet40 ModelNet10 T converge # views a Data (ModelNet40) # aug inst b aug Image based PANORAMA NN [25] 91.12 90.70 30 mins 1 a N Wang et al [27] 93.80 -20 hours 12 a N Pairwise [8] 90.70 92.80 -12 a N MVCNN [28] 90.10 --80 a N Geometry based Kd-Net depth 10 [9] 90.60 93. 30 clustering for the task of classification and takes 10 times longer to converge than our algorithm.…”
Section: Comparison With the State-of-the-art Methodsmentioning
confidence: 79%
“…Most of the missclassification comes from the similar category pairs like (table, desk ) and (night stand, dresser ). We also perform better than all the appearance based methods, except Wang et al [27] which performs a specialized view based ModelNet40 ModelNet10 T converge # views a Data (ModelNet40) # aug inst b aug Image based PANORAMA NN [25] 91.12 90.70 30 mins 1 a N Wang et al [27] 93.80 -20 hours 12 a N Pairwise [8] 90.70 92.80 -12 a N MVCNN [28] 90.10 --80 a N Geometry based Kd-Net depth 10 [9] 90.60 93. 30 clustering for the task of classification and takes 10 times longer to converge than our algorithm.…”
Section: Comparison With the State-of-the-art Methodsmentioning
confidence: 79%
“…The unordered point generation method (Fan, Su, and Guibas 2017) gets sparse point clouds which are limited to characterizing enough details, leading to large chamfer distances. The method proposed by Soltani et al(Soltani et al 2017) also obtains small distances since it generates points with many more (20) depth maps.…”
Section: Reconstruction Resultsmentioning
confidence: 99%
“…This architecture has been used to learn a continuous latent space for volumetric [17], [18], depthbaed [19], surface-based [20], and point-based [21] 3D reconstruction. In Wu et al [17], for example, the image encoder takes as input a 256×256 RGB image and outputs two 200-dimensional vectors representing, respectively, the mean and the standard deviation of a Gaussian distribution in the 200-dimensional space.…”
Section: Continuous Latent Spacesmentioning
confidence: 99%
“…The approach uses a standard encoder-decoder and an additional network composed of three fully-connected layers to encode the viewpoint. Soltani et al [19] and Lin et al [69] followed the same approach but predicts the depth maps, along with their binary masks, from predefined view points. In both methods, the merging is performed in a post-processing step.…”
Section: Intermediatingmentioning
confidence: 99%