2014
DOI: 10.1016/j.jvcir.2013.11.003
|View full text |Cite
|
Sign up to set email alerts
|

Graph-based approach for human action recognition using spatio-temporal features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
7
0

Year Published

2014
2014
2019
2019

Publication Types

Select...
6
3
1

Relationship

5
5

Authors

Journals

citations
Cited by 55 publications
(7 citation statements)
references
References 28 publications
0
7
0
Order By: Relevance
“…Having no previous knowledge about the location of the person in each video frame, the human action in a video stream can be recovered from a great number of local descriptors extracted from the video frames (Sekma et al, 2013), (Dammak et al, 2012), , (Sekma et al, 2014). Local descriptors, coupled with the bag-of-words (BOW) encoding method (Sivic and Zisserman, 2003) (Mejdoub et al, 2008) (Mejdoub et al, 2007) have recently become a very popular video representation (Ben Aoun et al, 2014), (Knopp et al, 2010), (Laptev et al, 2008), (Wang et al, 2009), (Alexander et al, 2008), (Wang et al, 2011), (Raptis and Soatto, 2010), (Pyry et al, 2010), (Jiang et al, 2012) and (Jain et al, 2013). The BOW uses a codebook to create a representation based on the visual content of a video, where the codebook is a set of visual words that represents the distribution of features of all the video.…”
Section: Intoductionmentioning
confidence: 99%
“…Having no previous knowledge about the location of the person in each video frame, the human action in a video stream can be recovered from a great number of local descriptors extracted from the video frames (Sekma et al, 2013), (Dammak et al, 2012), , (Sekma et al, 2014). Local descriptors, coupled with the bag-of-words (BOW) encoding method (Sivic and Zisserman, 2003) (Mejdoub et al, 2008) (Mejdoub et al, 2007) have recently become a very popular video representation (Ben Aoun et al, 2014), (Knopp et al, 2010), (Laptev et al, 2008), (Wang et al, 2009), (Alexander et al, 2008), (Wang et al, 2011), (Raptis and Soatto, 2010), (Pyry et al, 2010), (Jiang et al, 2012) and (Jain et al, 2013). The BOW uses a codebook to create a representation based on the visual content of a video, where the codebook is a set of visual words that represents the distribution of features of all the video.…”
Section: Intoductionmentioning
confidence: 99%
“…It aims to group image pixels into semantically meaningful regions. It has been used for many applications such as video action and event recognition [Wal10a,Ben11a,Ben14a,Ben14b,Mej15a], image search engines [Wan14a,Ben10a], augmented reality [Alh17a], image and video coding [Ben11b,Ben12a], Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee.…”
Section: Introductionmentioning
confidence: 99%
“…Indeed, actual person re-ID works are divided according to two main categories: shallow and deep methods. Shallow methods are specifically based on the appearance hand-crafted features [1,2,3,4,5,6,7,12,13,14,15]. In this context, two types of features are distinguished: low-level as well as mid-level ones.…”
Section: Introductionmentioning
confidence: 99%