2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00552
|View full text |Cite
|
Sign up to set email alerts
|

SelfRecon: Self Reconstruction Your Digital Avatar from Monocular Video

Abstract: Figure 1. We propose MultiPly, a novel framework to reconstruct multiple people in 3D from in-the-wild monocular videos. Our method can recover the complete 3D human with high-fidelity shape and appearance, even in scenarios involving occlusions and close interactions.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0
3

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 91 publications
(26 citation statements)
references
References 39 publications
(64 reference statements)
0
23
0
3
Order By: Relevance
“…To break through the limited representation ability of explicit mesh based digital human representation, many works adopt the implicit representation to improve the model capacity and visual quality [Gafni et al 2021;Jiang et al 2022;Yenamandra et al 2021;Zhuang et al 2021]. i3DMM [Yenamandra et al 2021] is the first neural implicit function based 3D morphable model of full heads.…”
Section: Related Workmentioning
confidence: 99%
“…To break through the limited representation ability of explicit mesh based digital human representation, many works adopt the implicit representation to improve the model capacity and visual quality [Gafni et al 2021;Jiang et al 2022;Yenamandra et al 2021;Zhuang et al 2021]. i3DMM [Yenamandra et al 2021] is the first neural implicit function based 3D morphable model of full heads.…”
Section: Related Workmentioning
confidence: 99%
“…ICON [37] infers a 3D clothed human meshes from a color image by utilizing a body-guided normal prediction model and a local-featurebased implicit 3D representation conditioned on SMPL(-X). SelfRecon [11] represents the human body as a template mesh and SDF in canonical space and utilizes a deformation field consisting of rigid forward LBS deformation and small non-rigid deformation to generate correspondences. Given monocular self-rotation RGB inputs, these methods are ca-pable of generating meshes of clothed humans.…”
Section: Related Workmentioning
confidence: 99%
“…Although INGP [22] is capable of converging in a short time and rendering high-fidelity novel view images, it is only suitable for static scenes and needs dense multi-view inputs. To aggregate the corresponding point information of different frames, we introduce a surface-relative representation relative to the surface point cloud of the human body (in practice, SMPL [19] or the mesh obtained by Sel-fRecon [11]). We also adopt a multi-resolution hash encoding [22] to speed up training.…”
Section: Modelmentioning
confidence: 99%
See 2 more Smart Citations