2021
DOI: 10.3390/s21113769
|View full text |Cite
|
Sign up to set email alerts
|

A Baseline for Cross-Database 3D Human Pose Estimation

Abstract: Vision-based 3D human pose estimation approaches are typically evaluated on datasets that are limited in diversity regarding many factors, e.g., subjects, poses, cameras, and lighting. However, for real-life applications, it would be desirable to create systems that work under arbitrary conditions (“in-the-wild”). To advance towards this goal, we investigated the commonly used datasets HumanEva-I, Human3.6M, and Panoptic Studio, discussed their biases (that is, their limitations in diversity), and illustrated … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(9 citation statements)
references
References 95 publications
(119 reference statements)
0
9
0
Order By: Relevance
“…Max RMSE errors were also lower in WF. According to previous findings [18,29,30], there is a bias in pose estimation algorithms regarding the position and distances of the used cameras and currently available open-source training datasets that were never designed with biomechanical applications in mind. We can conclude that the errors observed in the markerless system related to directional differences are primarily due to less effective tracking when participants have their backs turned toward the iPhone cameras and walk away from the cameras.…”
Section: Discussionmentioning
confidence: 99%
“…Max RMSE errors were also lower in WF. According to previous findings [18,29,30], there is a bias in pose estimation algorithms regarding the position and distances of the used cameras and currently available open-source training datasets that were never designed with biomechanical applications in mind. We can conclude that the errors observed in the markerless system related to directional differences are primarily due to less effective tracking when participants have their backs turned toward the iPhone cameras and walk away from the cameras.…”
Section: Discussionmentioning
confidence: 99%
“…Open-VICO is the ideal environment to crosscompare and fuse the results of these systems offering the footprint of a bottleneck custom ROS message for joint position and name harmonization compatible with the human landmarks described in sectio II-B. In [26] there are some guidelines on how to perform this harmonization procedure among different training datasets (e.g., COCO) typically used in deep learning-based skeleton tracking systems. Open-VICO's structure allows the user to easily append additional tracking algorithms to the default list, enriching the possible comparison combinations.…”
Section: Skeleton Tracking Methodsmentioning
confidence: 99%
“…HPE algorithms sometimes output incorrect keypoint coordinates, and researchers are still working on solving this issue [34]. Although there are some open datasets for researchers to develop action-recognizing models, most of these datasets are used for learning everyday actions, such as handshaking, talking, standing, and sitting, and do not apply to infant head lag studies [42]. In addition to calling for more open datasets in this field, further studies could explore the potential of the generative adversarial network (GAN) to influence the technology-assisted diagnosis of developmental delays in infants [21].…”
Section: Limitationsmentioning
confidence: 99%