2017
DOI: 10.1016/j.jvcir.2017.09.007
|View full text |Cite
|
Sign up to set email alerts
|

A novel recurrent hybrid network for feature fusion in action recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 23 publications
(11 citation statements)
references
References 36 publications
0
11
0
Order By: Relevance
“…On the other hand, feature concatenation often integrates features before classification by direct concatenation of the features extracted from individual modalities. Yu et al concatenated semantic features, long-term temporal features, and short-term temporal features of a video [43]. Ji et al concatenated object features, motion features and scene features from videos for linear classification [44].…”
Section: Fusion Methodsmentioning
confidence: 99%
“…On the other hand, feature concatenation often integrates features before classification by direct concatenation of the features extracted from individual modalities. Yu et al concatenated semantic features, long-term temporal features, and short-term temporal features of a video [43]. Ji et al concatenated object features, motion features and scene features from videos for linear classification [44].…”
Section: Fusion Methodsmentioning
confidence: 99%
“…To effectively learn spatiotemporal features, they apply a residual connection from the spatial stream to the temporal stream. Meantime, inspired by the success of recurrent neural networks in sequential information modeling [73]- [76], many researchers [42], [44], [45], [48], [77], [78] propose LSTM model for action recognition. Ng et al [44] and Donahue et al [45] extracted framelevel features of video by using CNNs model, and train LSTM with the frame-level feature for direct video-level prediction.…”
Section: B Deep Learning For Action Recognitionmentioning
confidence: 99%
“…Srivastava et al [48] proposed an approach for learning the sequence information in unsupervised settings by using LSTM architecture. To mitigate the overfitting problem, Yu et al [42] proposed a single-layer LSTM frameworks for learning long-term motion features. To learn spatio-temporal information, Zhang et al [79] proposed multi-level recurrent residual networks to produce complementary representations for action recognition.…”
Section: B Deep Learning For Action Recognitionmentioning
confidence: 99%
See 2 more Smart Citations