2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015
DOI: 10.1109/cvpr.2015.7298698
|View full text |Cite
|
Sign up to set email alerts
|

ActivityNet: A large-scale video benchmark for human activity understanding

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

3
879
0
1

Year Published

2015
2015
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 1,902 publications
(883 citation statements)
references
References 32 publications
3
879
0
1
Order By: Relevance
“…We evaluate our formulation on a large-scale, realistic activity dataset: ActivityNet [4]. Using our proposed ranking losses in training significantly improves performance in both the activity detection and early activity detection tasks.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…We evaluate our formulation on a large-scale, realistic activity dataset: ActivityNet [4]. Using our proposed ranking losses in training significantly improves performance in both the activity detection and early activity detection tasks.…”
Section: Methodsmentioning
confidence: 99%
“…The ActivityNet [4] dataset comprises 28K videos of 203 activity categories collected from YouTube. Fig.…”
Section: Datasetmentioning
confidence: 99%
See 1 more Smart Citation
“…We report results on the 213 test videos with temporal annotations. To study the generalization capability of our model across datasets, we also test on the validation set of the ActivityNet benchmark (release 1.2) [3], which comprises 76 hours of video and 100 action classes. No fine tuning is done on this benchmark.…”
Section: Methodsmentioning
confidence: 99%
“…However, it was recently shown that the temporal footprint of some methods can be as accurate as sampling temporal proposals uniformly in the video [4]. Moreover, these methods evaluate their performance on simple or repetitive actions in short video clips, which makes it difficult to gauge their scalability to large collections of video sequences containing more challenging activities [18,3]. Given the current state-of-the-art of spatio-temporal action proposals, it is worth exploring how only temporal action proposals can contribute to the semantic analysis of videos.…”
Section: Introductionmentioning
confidence: 99%