2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.01031
|View full text |Cite
|
Sign up to set email alerts
|

Model-based 3D Hand Reconstruction via Self-Supervised Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 86 publications
(36 citation statements)
references
References 32 publications
0
21
0
Order By: Relevance
“…Baek et al [52] address the challenging problem of hand-object interaction scenarios and combine a generative adversarial network and mesh renderer for guidance. Chen et al [3] employ an off-the-shell 2D pose detector [5] as a weaker 2D supervision, compared to human annotated 2D keypoints. By taking advantage of hand detector, the model can be trained on a wilder range of images without human label.…”
Section: Weakly-supervised Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Baek et al [52] address the challenging problem of hand-object interaction scenarios and combine a generative adversarial network and mesh renderer for guidance. Chen et al [3] employ an off-the-shell 2D pose detector [5] as a weaker 2D supervision, compared to human annotated 2D keypoints. By taking advantage of hand detector, the model can be trained on a wilder range of images without human label.…”
Section: Weakly-supervised Methodsmentioning
confidence: 99%
“…R ECENTLY, a surge of research efforts [1] [2] [3] have been devoted to 3D hand reconstruction. In contrast to the conventional approaches relying on RGB-D sensor [4] or multiple view geometry [5], recovering 3D hand pose and its shape from single color image is more challenging due to the ambiguities in depth and scale.…”
Section: Introductionmentioning
confidence: 99%
“…In recent years, scene understanding technologies for use in autonomous driving [19], virtual reality [20] and augmented reality [21] have developed rapidly. As the basic task of scene understanding, semantic segmentation technology based on pixel-by-pixel classification has been widely studied [22][23][24].…”
Section: Semantic Segmentation Based On Deep Learningmentioning
confidence: 99%
“…Popular hand mesh estimation methods can be divided into five types, whose core ideas are based on the parametric model, voxel representation, implicit function, UV map, and vertex position, respectively. Model-based approaches [86,90,80,29,85,4,1,93,12,88,6,2,39,52,82,83] typically use MANO [65] as the parametric model, which factorizes a hand mesh into coefficients of shape and pose. This pipeline, however, is not suitable for usage with a lightweight network because the coefficient estimation is a highly abstract problem that ignores spatial correlations [22].…”
Section: Related Workmentioning
confidence: 99%