2017 IEEE International Conference on Computer Vision (ICCV) 2017
DOI: 10.1109/iccv.2017.39
|View full text |Cite
|
Sign up to set email alerts
|

Encouraging LSTMs to Anticipate Actions Very Early

Abstract: In contrast to the widely studied problem of recognizing an action given a complete sequence, action anticipation aims to identify the action from only partially available videos. As such, it is therefore key to the success of computer vision applications requiring to react as early as possible, such as autonomous navigation. In this paper, we propose a new action anticipation method that achieves high prediction accuracy even in the presence of a very small percentage of a video sequence. To this end, we deve… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
206
0

Year Published

2018
2018
2019
2019

Publication Types

Select...
5
1

Relationship

2
4

Authors

Journals

citations
Cited by 158 publications
(206 citation statements)
references
References 52 publications
0
206
0
Order By: Relevance
“…Earliest Latest Context-aware + loss in [21] 30.6 71.1 Context-aware + loss in [34] 22.6 73.1 Multi stage LSTM [3] 80.5 83.4 Proposed 84.2 85.6 Table 1. Action anticipation results for UCF101 considering the 'Earliest' 20% of frames and 'Latest' 50% of frames.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…Earliest Latest Context-aware + loss in [21] 30.6 71.1 Context-aware + loss in [34] 22.6 73.1 Multi stage LSTM [3] 80.5 83.4 Proposed 84.2 85.6 Table 1. Action anticipation results for UCF101 considering the 'Earliest' 20% of frames and 'Latest' 50% of frames.…”
Section: Methodsmentioning
confidence: 99%
“…Lee et al [30] proposed a human activity representation method, termed sub-volume co-occurrence matrix, and developed a method to predict partially observed actions with the aid of a pretrained CNN. The deep network approach of Aliakbarian et al [3] used a multi-stage LSTM architecture that incorporates context-aware and action-aware features to predict classes as early as possible. The CNN based action anticipation model of [40] predicts the most plausible future motion, and was improved via an effective loss function based on dynamic and classification losses.…”
Section: Previous Workmentioning
confidence: 99%
See 3 more Smart Citations