2021
DOI: 10.1016/j.neucom.2021.01.045
|View full text |Cite
|
Sign up to set email alerts
|

Spatial-aware stacked regression network for real-time 3D hand pose estimation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
27
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 27 publications
(27 citation statements)
references
References 22 publications
0
27
0
Order By: Relevance
“…Nevertheless, due to the adoption of a complex ResNet-50-based backbone network, A2J requires a higher computational complexity than the proposed method as described in Table 4. As compared to SRN [14] which shows the best performance in both 2D and 3D input data-based methods, our method achieves comparable results on ICVL and MSRA datasets. For NYU dataset, our method is inferior to SRN on accuracy but shows better running efficiency.…”
Section: Comparision With State-of-the-art Methodsmentioning
confidence: 90%
See 3 more Smart Citations
“…Nevertheless, due to the adoption of a complex ResNet-50-based backbone network, A2J requires a higher computational complexity than the proposed method as described in Table 4. As compared to SRN [14] which shows the best performance in both 2D and 3D input data-based methods, our method achieves comparable results on ICVL and MSRA datasets. For NYU dataset, our method is inferior to SRN on accuracy but shows better running efficiency.…”
Section: Comparision With State-of-the-art Methodsmentioning
confidence: 90%
“…We compared the proposed network on three public 3D hand pose datasets with the most recently proposed methods using 2D depth maps as an input, including DISCO [35], Deep-Prior [36], its improved version DeepPrior++ [29], Feedback [16], Multi-view CNNs [34], REN-4 × 6 × 6 [38], REN-9 × 6 × 6 [39], Pose-REN [40], Generalized [41], Global2Local [10], CrossingNets [37], HBE [12], Cross-InfoNet [9], A2J [15], and SRN [14] as well as methods using 3D inputs, including 3D CNN [3], SHPR-Net [6], 3D DenseNet [4], HandPointNet [7], Point-to-Point [8], and V2V-PoseNet [5]. The average 3D distance error per joint and percentage of success frames over different error thresholds are respectively shown in Table 3 and Fig.…”
Section: Comparision With State-of-the-art Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…The "force" will drive a 2D hand skeleton to "fold" into the 3D joint coordinates representing the hand pose. adopted to perform direct hand depth image processing [40,10,14,30,3]. However, 2D CNNs cannot fully take advantage of 3D spatial information of the depth image, which is essential for achieving high accuracy.…”
Section: Introductionmentioning
confidence: 99%