2021
DOI: 10.48550/arxiv.2104.03670
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

3D Shape Generation and Completion through Point-Voxel Diffusion

Abstract: Figure 1: The proposed Point-Voxel Diffusion (PVD) is a new framework for generative modeling of 3D shapes. Left: tables, cars, and planes generated by our PVD. It learns to sample from a Gaussian prior and to progressively remove noise to obtain sharp shapes. Right: two possible shapes completed from a real RGB-D image, each visualized in input and canonical views.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 28 publications
0
7
0
Order By: Relevance
“…Diffusion-based generative models Our work lays a theoretical foundation for , which recognizes that conditional denoising score matching (Song & Ermon, 2019 and discrete-time diffusion-based generative models (Sohl-Dickstein et al, 2015;Ho et al, 2020) can be viewed as learning to revert an inference process (using the plug-in reverse SDE). This line of work has been successfully applied to modeling high dimensional natural images (Dhariwal & Nichol, 2021;Saharia et al, 2021), audio (Kong et al, 2020), 3D point cloud (Cai et al, 2020;Zhou et al, 2021), and discrete data (Hoogeboom et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…Diffusion-based generative models Our work lays a theoretical foundation for , which recognizes that conditional denoising score matching (Song & Ermon, 2019 and discrete-time diffusion-based generative models (Sohl-Dickstein et al, 2015;Ho et al, 2020) can be viewed as learning to revert an inference process (using the plug-in reverse SDE). This line of work has been successfully applied to modeling high dimensional natural images (Dhariwal & Nichol, 2021;Saharia et al, 2021), audio (Kong et al, 2020), 3D point cloud (Cai et al, 2020;Zhou et al, 2021), and discrete data (Hoogeboom et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…We utilize free-form deformation (FFD) (Sederberg & Parry, 1986) and radial basis function (RBF)-based deformation (Forti & Rozza, 2014) for non-linear transformations. Such deformations are also common in VR/AR games and point clouds from generative models (GAN) (Li et al, 2018a;Zhou et al, 2021). Specifically, we use multi quadratic (ϕ(x) = √ x 2 + r 2 ) and inverse multi quadratic splines (ϕ(x) = (x 2 + r 2 ) − 1 2 ) as the representative RBFs to cover a wide range of deformation types.…”
Section: Transformation Corruptions Patternsmentioning
confidence: 99%
“…We show results under trajectories of different number of timesteps K. We select the minimum K such that analytic-DPM can outperform the baselines with full timesteps and underline the corresponding results. (Chen et al, 2020;Kong et al, 2020;Popov et al, 2021;Lam et al, 2021), controllable generation (Choi et al, 2021;Sinha et al, 2021), image super-resolution (Saharia et al, 2021;, image-to-image translation (Sasaki et al, 2021), shape generation (Zhou et al, 2021) and time series forecasting (Rasul et al, 2021).…”
Section: Related Workmentioning
confidence: 99%