2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI) 2018
DOI: 10.1109/sibgrapi.2018.00064
|View full text |Cite
|
Sign up to set email alerts
|

Bag of Attributes for Video Event Retrieval

Abstract: In this paper, we present the Bag-of-Attributes (BoA) model for video representation aiming at video event retrieval. The BoA model is based on a semantic feature space for representing videos, resulting in high-level video feature vectors. For creating a semantic space, i.e., the attribute space, we can train a classifier using a labeled image dataset, obtaining a classification model that can be understood as a high-level codebook. This model is used to map low-level frame vectors into high-level vectors (e.… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 38 publications
0
1
0
Order By: Relevance
“…Numerous deep learning methods for human action recognition have appeared in the literature [2], [7]- [11]. In most of them, a video is parsed frame by frame with convolutional neural networks (CNNs) designed for images [12], [13]. Other methods process videos as image sequences using 2D CNNs, 3D CNNs, or recurrent neural networks (RNNs) [14]- [16].…”
Section: Introductionmentioning
confidence: 99%
“…Numerous deep learning methods for human action recognition have appeared in the literature [2], [7]- [11]. In most of them, a video is parsed frame by frame with convolutional neural networks (CNNs) designed for images [12], [13]. Other methods process videos as image sequences using 2D CNNs, 3D CNNs, or recurrent neural networks (RNNs) [14]- [16].…”
Section: Introductionmentioning
confidence: 99%