2016
DOI: 10.1007/s11263-016-0939-9
|View full text |Cite
|
Sign up to set email alerts
|

Behavior Discovery and Alignment of Articulated Object Classes from Unstructured Video

Abstract: We propose an automatic system for organizing the content of a collection of unstructured videos of an articulated object class (e.g., tiger, horse). By exploiting the recurring motion patterns of the class across videos, our system: (1) identifies its characteristic behaviors, and (2) recovers pixel-to-pixel alignments across different instances. Our system can be useful for organizing video collections for indexing and retrieval. Moreover, it can be a platform for learning the appearance or behaviors of obje… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 12 publications
(4 citation statements)
references
References 84 publications
(151 reference statements)
0
4
0
Order By: Relevance
“…We train the model with a dataset of video frames depicting full-body horses. First, we select the horse subset from the latest version of the TigDog dataset [9]. We use video frames for all the horse sequences in the dataset, discarding video frames showing partially visible horses.…”
Section: Datamentioning
confidence: 99%
See 1 more Smart Citation
“…We train the model with a dataset of video frames depicting full-body horses. First, we select the horse subset from the latest version of the TigDog dataset [9]. We use video frames for all the horse sequences in the dataset, discarding video frames showing partially visible horses.…”
Section: Datamentioning
confidence: 99%
“…We collect frames containing complete horses from about 60 videos, representing 47k frames (plus around 6k from [9]). Note that this dataset is relatively small compared to what is required for training human pose estimation models -our horse dataset is only 1.3% of the size of the Human3.6M dataset [13] and 3.6% of the size of the MPI-INF-3DHP dataset [21].…”
Section: Datamentioning
confidence: 99%
“…The existing action co-localization works have mainly focused on two scenarios, i.e., co-localization in pairs of videos [18,38,97] and weakly supervised action co-localization with video level labels [20,21,92,93,108,134,143]. Few of these works have considered a fully unconstrained scenario like us, i.e., the numbers and types of common actions are unknown in advance and each video may contain zero, one or several common actions.…”
Section: Thematic Action Discovery and Localization In Collections Of Videosmentioning
confidence: 99%
“…It proposes a feature to describe pairs of trajectories in order to better characterize articulated object motion. [21] extends [20] to also recovers the spatial alignment among different instances of the same behavior using a Thin Plate Spline deformation model. [147] proposes a method to mine recurring events in a collection of videos.…”
Section: Literature Reviewmentioning
confidence: 99%