2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00326
|View full text |Cite
|
Sign up to set email alerts
|

HOnnotate: A Method for 3D Annotation of Hand and Object Poses

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
308
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 299 publications
(308 citation statements)
references
References 54 publications
0
308
0
Order By: Relevance
“…Due to the lack of large scale 3D ground truth data, synthetic data has often been used for training [14,29,49]. Recently, instead of estimating the hand skeleton, recovering the pose and the surface of the hand has become popular using statistical hand models, e.g., the MANO model [74], that can represent a variety of hand shapes and poses [25,93,97]. Using the template derived from MANO, [41] show that it is also possible to regress hand meshes directly using mesh convolution.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Due to the lack of large scale 3D ground truth data, synthetic data has often been used for training [14,29,49]. Recently, instead of estimating the hand skeleton, recovering the pose and the surface of the hand has become popular using statistical hand models, e.g., the MANO model [74], that can represent a variety of hand shapes and poses [25,93,97]. Using the template derived from MANO, [41] show that it is also possible to regress hand meshes directly using mesh convolution.…”
Section: Related Workmentioning
confidence: 99%
“…Hand-object interaction is generated using a physics simulator, GraspIt [53], resulting in high-quality hand-object interaction. Due to the limited number of grasp types in the FHB dataset [19] and the HO-3D dataset [25], they are not suitable for training the generative model (see Appendix B). Instead, we use them to test the generalization ability of the generative model trained on the ObMan grasps.…”
Section: Datasetmentioning
confidence: 99%
See 1 more Smart Citation
“…We acquired RGB and depth frames and reconstruct the colorized overlapping point clouds for ground truth labeling while handling a real Colibri II drill. We choose this marker-less tracking approach over marker-based approaches to recover the ground truth hand and tool poses, since any markers attached to the tool or hand would be visible in the captured images and can introduce a bias for learning based methods [13].…”
Section: Real Data Generationmentioning
confidence: 99%
“…Since there is no other widely acknowledged benchmark dataset for depth data of hands interacting with objects from egocentric view, we chose other standard benchmark datasets: NYU, HO3D, and HANDS2017 to test the performance of our solution on data that are not exclusively egocentric. HO3D [68] is a dataset of RGB-D images of hands manipulating with objects from a 3 rd person view. Annotations of the pose of the hand and the object are available.…”
Section: ) Other Datasetsmentioning
confidence: 99%