2019
DOI: 10.1016/j.neucom.2019.05.058
|View full text |Cite
|
Sign up to set email alerts
|

Spatial-temporal pyramid based Convolutional Neural Network for action recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
16
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 51 publications
(16 citation statements)
references
References 17 publications
0
16
0
Order By: Relevance
“…Extracting local features from the shape can be done by solving a Poisson equation and finally using a template-based nearest neighbor classifier for classification. Xu et al and Zheng et al [15,16] tracked the main joints of the human body, used a parameterized method to represent the rotation and translation of each part of the human body, and used these parameters to express the actions. Poultangari et al [17] introduced cloud theory and particle swarm algorithm and used particle swarm optimization backpropagation neural network (PSO-BPNN) method for structural damage identification.…”
Section: Related Workmentioning
confidence: 99%
“…Extracting local features from the shape can be done by solving a Poisson equation and finally using a template-based nearest neighbor classifier for classification. Xu et al and Zheng et al [15,16] tracked the main joints of the human body, used a parameterized method to represent the rotation and translation of each part of the human body, and used these parameters to express the actions. Poultangari et al [17] introduced cloud theory and particle swarm algorithm and used particle swarm optimization backpropagation neural network (PSO-BPNN) method for structural damage identification.…”
Section: Related Workmentioning
confidence: 99%
“…[28] uses the relative position and relative speed to provide spatial information and dynamic information in skeletal action sequences, and build a bidirectional LSTM-CNN to fuse spatial-temporal information for skeleton-based action recognition. Spatial-Temporal Pyramid Network [29] (S-TPNet) extracts multiscale appearance features from different stages of the 2D CNN. It groups the frame features into different snippet-level features and uses fully-connected layers to reason about snippet relations.…”
Section: Relation Modelmentioning
confidence: 99%
“…Ke et al propose Frame Segmentation Network, improving mAP (mean average precision) and IoU (Intersection over Union) simultaneously [20]. As for action recognition, the S-TPNet proposed in [21] has a good performance. Its accuracy is around 74% on dataset HMDB51, and around 95% on dataset UCF101.…”
Section: Related Workmentioning
confidence: 99%