Proceedings of the 30th ACM International Conference on Multimedia 2022
DOI: 10.1145/3503161.3548546
|View full text |Cite
|
Sign up to set email alerts
|

PYSKL: Towards Good Practices for Skeleton Action Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
31
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 87 publications
(31 citation statements)
references
References 22 publications
0
31
0
Order By: Relevance
“…Theoretically, GCDAE can be integrated with any skeleton based action recognition model to improve its robustness to noisy data. We combine it with five SOTA recognition models mentioned in Section 2.1, that is, EfficientGCN [35], ST‐GCN++ [36], CTR‐GCN [21], AAGCN [13] and MS‐G3D [14] for experiments. As different methods utilize high‐order features in different ways, we use only the first‐order joint stream for simplicity and fair comparison.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…Theoretically, GCDAE can be integrated with any skeleton based action recognition model to improve its robustness to noisy data. We combine it with five SOTA recognition models mentioned in Section 2.1, that is, EfficientGCN [35], ST‐GCN++ [36], CTR‐GCN [21], AAGCN [13] and MS‐G3D [14] for experiments. As different methods utilize high‐order features in different ways, we use only the first‐order joint stream for simplicity and fair comparison.…”
Section: Methodsmentioning
confidence: 99%
“…We retrain the EfficientGCN‐B0 as their data preprocessing on CS benchmark is different from ours. For other four backbone models, pretrained weights provided by PYSKL [36] are directly used. Comparison of these backbones are shown in Table 1.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…The key to processing unorganized motion datasets is label acquisition. We demonstrate that an action recognition network [Duan et al 2022] pretrained on NTURGB+D 120 dataset [Liu et al 2020] can produce reliable skill labels. We test with raw unstructured motion clips from CMU Mocap dataset [CMU 2002] by finetuning the network on the aforementioned Composite Skills dataset, where we set the learning rate to be 0.1 with momentum being 0.9 and weight decay being 5×10 −4 .…”
Section: A2 Skill Labelmentioning
confidence: 99%