2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops) 2011
DOI: 10.1109/iccvw.2011.6130399
|View full text |Cite
|
Sign up to set email alerts
|

Probabilistic subspace-based learning of shape dynamics modes for multi-view action recognition

Abstract: We propose a human action recognition algorithm by capturing a compact signature of shape dynamics from multi-view videos. First, we compute R transforms and its temporal velocity on action silhouettes from multiple views to generate a robust low level representation of shape. The spatio-temporal shape dynamics across all the views is then captured by fusion of eigen and multiset partial least squares modes. This provides us a lightweight signature which is classified using a probabilistic subspace similarity … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2013
2013
2018
2018

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 13 publications
(8 citation statements)
references
References 13 publications
0
8
0
Order By: Relevance
“…A joint bag‐of‐words histogram might be built with local feature descriptors extracted from each one of the camera views (Wu et al ., ), but a better performance is reported when other fusion strategies are employed. Projections maximizing cross‐covariance have been learned to combine scriptR‐transform derivatives extracted from each camera view (Karthikeyan et al ., ). Two‐level linear discriminant analysis has been employed to learn silhouette projections maximizing action class separability (Iosifidis et al ., ).…”
Section: Related Workmentioning
confidence: 97%
“…A joint bag‐of‐words histogram might be built with local feature descriptors extracted from each one of the camera views (Wu et al ., ), but a better performance is reported when other fusion strategies are employed. Projections maximizing cross‐covariance have been learned to combine scriptR‐transform derivatives extracted from each camera view (Karthikeyan et al ., ). Two‐level linear discriminant analysis has been employed to learn silhouette projections maximizing action class separability (Iosifidis et al ., ).…”
Section: Related Workmentioning
confidence: 97%
“…In [43], Delaitre et al proposed a statistical model which integrates order-less person-object interaction responses to recognize actions in still images by calculating spatial co-occurrences of individual body parts and objects. To achieve the property of view invariance, Karthikeyan et al [8] computed the R transforms on action silhouettes from multiple views and then proposed a probabilistic subspace similarity technique to recognize actions by learning their inter-action and intra-action models. However, when recognizing an action, the scheme requires that all action sequences captured at different views should be collected together simultaneously.…”
Section: Related Workmentioning
confidence: 99%
“…There have been many approaches [8,9,[13][14][15][16][17][18]42,44] proposed in the literature for video-based human movement analysis. For example, in [52], Fengjun and Nevatia used a posture-based scheme to convert frames to strings from which actions were then classified into different types using the action nets.…”
Section: Related Workmentioning
confidence: 99%
“…Furthermore, Cuzzolin [23] built different bilinear HMMs to analyze human gaits from multiple views. Karthikeyan, Gaur, and Manjunath [24] computed R transforms on action silhouettes and then proposed a probabilistic subspace similarity technique to recognize actions from multiple views by learning their inter-action and intra-action models. However, this scheme requires all action sequences from different views must be collected together before recognize an action.…”
Section: Introductionmentioning
confidence: 99%