2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00175
|View full text |Cite
|
Sign up to set email alerts
|

Multi-View Neural Human Rendering

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
83
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 99 publications
(84 citation statements)
references
References 27 publications
1
83
0
Order By: Relevance
“…The free-view results, however, are vulnerable to occlusions and suffer from limited view interpolation along with the dense captured views, leading to uncanny texture details due to view blending. The recent neural rendering techniques bring huge potential for compelling photo-realistic free-viewpoint video generation via neural view blending [Meshry et al 2019a;Thies et al 2018] or neural scene modeling [Bemana et al 2020;Mildenhall et al 2020a;Wu et al 2020b]. Open4D [Bansal et al 2020] generates a free-viewpoint video enabling occlusion removal and time-freezing effects using around 15 mobile cameras, which is similar to our work.…”
Section: Introductionsupporting
confidence: 74%
See 1 more Smart Citation
“…The free-view results, however, are vulnerable to occlusions and suffer from limited view interpolation along with the dense captured views, leading to uncanny texture details due to view blending. The recent neural rendering techniques bring huge potential for compelling photo-realistic free-viewpoint video generation via neural view blending [Meshry et al 2019a;Thies et al 2018] or neural scene modeling [Bemana et al 2020;Mildenhall et al 2020a;Wu et al 2020b]. Open4D [Bansal et al 2020] generates a free-viewpoint video enabling occlusion removal and time-freezing effects using around 15 mobile cameras, which is similar to our work.…”
Section: Introductionsupporting
confidence: 74%
“…However, this method suffers from limited free-viewing range and fragile dynamic motion modeling. For reconstructing neural scenes, various data representations have been explored, such as point-clouds [Aliev et al 2020;Wu et al 2020b], voxels [Lombardi et al 2019;Sitzmann et al 2019a] or implicit representations [Mildenhall et al 2020a;Sitzmann et al 2019b;Suo et al 2021]. Researchers [Jin et al 2018;Kwon et al 2020] also utilize the underlying latent geometry for novel view synthesis of human performance in the encoder-decoder manner, which still suffers from limited representation ability of a single latent code for complex human inferior texture output.…”
Section: Related Workmentioning
confidence: 99%
“…The results in Table 1 verify our advanced generalization capacity on the unseen scenarios. We also achieve competitive fitting performance on the training frames, even comparable to the per-scene optimization methods [28,37,39].…”
Section: Synthesis Performance Analysismentioning
confidence: 93%
“…Recently, researchers adopt neural networks to represent the shape and appearance of scenes. These representations, such as voxels [19,21,26,33], point clouds [1,39], textured meshes [14,16,18,40] and multi-plane images [7,44] are learned from 2D images via differentiable renderers. Though with impressive results, they are hard to scale to higher resolution due to innate cubic memory complexity.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation