2018
DOI: 10.1007/978-3-030-01270-0_21
|View full text |Cite
|
Sign up to set email alerts
|

Deep Volumetric Video From Very Sparse Multi-view Performance Capture

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
95
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 116 publications
(95 citation statements)
references
References 72 publications
0
95
0
Order By: Relevance
“…However, due to possible inconsistency between the synthesized silhouettes, the subtraction operation of visual hull tends to excessively erode the reconstructed mesh. To further improve the output quality, we adopt a deep visual hull algorithm similar to Huang et al [20] with a greedy view sampling strategy so that the reconstruction results account for domain-specific clothed human body priors (Sec. 3.2).…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…However, due to possible inconsistency between the synthesized silhouettes, the subtraction operation of visual hull tends to excessively erode the reconstructed mesh. To further improve the output quality, we adopt a deep visual hull algorithm similar to Huang et al [20] with a greedy view sampling strategy so that the reconstruction results account for domain-specific clothed human body priors (Sec. 3.2).…”
Section: Methodsmentioning
confidence: 99%
“…In particular, we use a network structure based on [20]. At a high level, Huang et al [20] propose to map 2D images to a 3D volumetric field through a multi-view convolutional neural network. The 3D field encodes the probabilistic distribution of 3D points on the captured surface.…”
Section: Deep Visual Hull Predictionmentioning
confidence: 99%
See 1 more Smart Citation
“…In order to replicate the success of 2D convolutional neural network to 3D domain, various forms of 3D representations have been explored. As a natural extension of 2D pixels, volumetric representation has been widely used in recent works on 3D reconstruction [29,30,34,6,13,31,36,10,33] due to its simplicity of implementation and compatibility with convolutional neural network. However, deep voxel generators are constrained by its resolution due to the data sparsity and computation cost of 3D convolution.…”
Section: Related Workmentioning
confidence: 99%
“…A large body of work exists to extract human representations from multiple input views or sensors, of which some recently use deep learning to extract 3D human representations [8,13,21]. While they intrinsically aren't designed to deal with monocular input as proposed, multi-view methods usually yield more complete and higher precision results as soon as several viewpoints are available, a useful feature we leverage for creating the 3D HUMANS dataset.…”
Section: Related Workmentioning
confidence: 99%