2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.01126
|View full text |Cite
|
Sign up to set email alerts
|

Revitalizing Optimization for 3D Human Pose and Shape Estimation: A Sparse Constrained Formulation

Abstract: We propose a novel sparse constrained formulation and from it derive a real-time optimization method for 3D human pose and shape estimation. Our optimization method is orders of magnitude faster (avg. 4 ms convergence) than existing optimization methods, while being mathematically equivalent to their dense unconstrained formulation. We achieve this by exploiting the underlying sparsity and constraints of our formulation to efficiently compute the Gauss-Newton direction. We show that this computation scales lin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 22 publications
(11 citation statements)
references
References 36 publications
0
11
0
Order By: Relevance
“…Output Type a) Template paramters: [16], [17], [18], [19], [20], [21], [25], [37], [94], [97], [98], [99], [101], [102], [103], [104], [105], [106], [111], [112], [125], [129], [135], [140] b) 3D vertex coordinates: GraphCMR [117], Pose2Mesh [132], I2L-MeshNet [118], PC-HMR [121], METRO [120], Graphormer [119] c) Voxels: BodyNet [115], DeepHuman [116] d) UV position maps: DenseBody [122], DecoMR [123], Zhang et al [124] e) Probabilistic outputs: Biggs et al [125], Sengupta et al [127], [128], ProHMR [126] Intermediate/ Proxy Representation a) Silhouettes: Pavlakos et al [17], STRAPS [103], Skeleton2Mesh [98] b) Segmentations: NBF [18], Rueegg et al [129], STRAPS [103], Zanfir et al [105], HUND…”
Section: Framebased Single Personmentioning
confidence: 99%
See 1 more Smart Citation
“…Output Type a) Template paramters: [16], [17], [18], [19], [20], [21], [25], [37], [94], [97], [98], [99], [101], [102], [103], [104], [105], [106], [111], [112], [125], [129], [135], [140] b) 3D vertex coordinates: GraphCMR [117], Pose2Mesh [132], I2L-MeshNet [118], PC-HMR [121], METRO [120], Graphormer [119] c) Voxels: BodyNet [115], DeepHuman [116] d) UV position maps: DenseBody [122], DecoMR [123], Zhang et al [124] e) Probabilistic outputs: Biggs et al [125], Sengupta et al [127], [128], ProHMR [126] Intermediate/ Proxy Representation a) Silhouettes: Pavlakos et al [17], STRAPS [103], Skeleton2Mesh [98] b) Segmentations: NBF [18], Rueegg et al [129], STRAPS [103], Zanfir et al [105], HUND…”
Section: Framebased Single Personmentioning
confidence: 99%
“…Inspired by this, Biggs et al [125] also adopt the Real-NVP architecture. Fan et al [140] design a normalizing flow using fully-connected layers. The GHUM/GHUML models [81] rely on normalizing flows to represent skeleton kinematics.…”
Section: Pose Prior and Shape Priormentioning
confidence: 99%
“…Furthermore, adapting the optimizer to run in real-time is a nontrivial operation, due to the cubic complexity of the popular Levenberg-Marquardt algorithm [26,38,44]. The most common and practical way to speedup the optimization is to utilize the sparsity of the problem or make certain assumptions to simplify it [17]. Learned optimizers promise to overcome these issues, by learning the parametric model priors directly from the data and to take more aggressive steps, thus converging in fewer iterations.…”
Section: Related Workmentioning
confidence: 99%
“…Recent years have witnessed significant development of marker-less motion capture, which promotes a wide variety of applications ranging from character animation to human-computer interaction, personal well-being, and human behavior understanding. Extensive existing works can kinematically capture accurate human pose from monocular videos and images via network regression [23,26,27,72,73] or optimization [6,42,45,55]. However, they are often hard to leverage in real-world systems due to a series of artifacts that are not satisfied biomechanical and physical plausibility (e.g., jitter and floor penetration).…”
Section: Introductionmentioning
confidence: 99%