2020
DOI: 10.1016/j.patcog.2020.107293
|View full text |Cite
|
Sign up to set email alerts
|

Learning shape and motion representations for view invariant skeleton-based action recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
16
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 41 publications
(16 citation statements)
references
References 15 publications
0
16
0
Order By: Relevance
“…At present, the deep learning methods for skeleton-based action recognition are mainly divided into three categories: (1) The RNN-based methods [36,30,29,31] is based on the natural time properties of the skeleton sequence, and then modeled by Long Short Term Memory (LSTM), Gated Recurrent Unit (GRU), etc. ; (2) The CNNbased methods [19,12,42,17] usually convert the skeleton sequence into the pseudoimages using specific transformation rules, and model it with efficient image classification networks. The CNN-based methods usually combined with RNN-based methods to model the temporal information of skeleton sequence.…”
Section: Skeleton-based Action Recognitionmentioning
confidence: 99%
“…At present, the deep learning methods for skeleton-based action recognition are mainly divided into three categories: (1) The RNN-based methods [36,30,29,31] is based on the natural time properties of the skeleton sequence, and then modeled by Long Short Term Memory (LSTM), Gated Recurrent Unit (GRU), etc. ; (2) The CNNbased methods [19,12,42,17] usually convert the skeleton sequence into the pseudoimages using specific transformation rules, and model it with efficient image classification networks. The CNN-based methods usually combined with RNN-based methods to model the temporal information of skeleton sequence.…”
Section: Skeleton-based Action Recognitionmentioning
confidence: 99%
“…Many researchers have spent their valuable time preparing some challenging datasets based on the skeleton of the human body for further experiments including NTU, MSR Action3D, Berkeley MHAD, HDM05, and UTD multimodal human action dataset (UTD-MHAD) datasets. Several ideas developed based on skeleton data [16][17][18][19][20][21][22][23][24][25][26][27][28][29] are used to separate action classes. In [16], J. Imran et al evaluated a method based on skeleton augmented data of 3D skeleton joints information using 3D transformations and designed a RNN-based BiGRU for the purpose of classification.…”
Section: Skeleton-based Action Recognitionmentioning
confidence: 99%
“…For dynamic skeleton sequences, they also proposed a temporal stack learning network. Li et al [26] represented skeleton sequences as a subset of geometric algebra to extract temporal and spatial features such as shape and motion. They also designed a rotor based view transformation method and spatio temporal view-invariant model to eliminate the effect of viewpoint variation and to combine skeleton joints and bones to capture spatial configuration and temporal dynamics, respectively.…”
Section: Skeleton-based Action Recognitionmentioning
confidence: 99%
“…We could still recognize the same action for both actors through the movement of the actors' feet and elbows according to key point shifts. Previous methods used in the skeleton-based action recognition task [31,32] have proven the effectiveness of key points and their shifts for recognizing human actions, where the key points and their correspondence across frames are provided as inputs. Skeleton data outlines the key points of actors in the video, suppressing the influence of trivial environment information during the feature extraction process.…”
Section: Fully-supervised Learningmentioning
confidence: 99%
“…Key points and their displacements have been used mainly in skeleton-based action recognition [10,31,32,59]. Most of the skeleton-based methods take skeleton data as the input, which is generated by devices [33] or pose estimation algorithms [35] in the form of 2D or 3D coordinates.…”
Section: Skeleton-based Action Recognitionmentioning
confidence: 99%