2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015
DOI: 10.1109/cvpr.2015.7298981
|View full text |Cite
|
Sign up to set email alerts
|

Video co-summarization: Video summarization by visual co-occurrence

Abstract: We present video co-summarization, a novel perspective to video summarization that exploits visual co-occurrence across multiple videos. Motivated by the observation that important visual concepts tend to appear repeatedly across videos of the same topic, we propose to summarize a video by finding shots that co-occur most frequently across videos collected using a topic keyword. The main technical challenge is dealing with the sparsity of co-occurring patterns, out of hundreds to possibly thousands of irreleva… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
179
0

Year Published

2015
2015
2019
2019

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 216 publications
(179 citation statements)
references
References 34 publications
0
179
0
Order By: Relevance
“…As in [5], the frames are initially clustered in sequential order, as consecutive frames are similar and more probably to be allocated to the same cluster. Other works construct more comprehensive models based on the idea of clustering [3,22]. As in [22], the video is transformed into an undirected graph, and the summary is generated by partitioning this graph into clusters.…”
Section: Related Workmentioning
confidence: 99%
“…As in [5], the frames are initially clustered in sequential order, as consecutive frames are similar and more probably to be allocated to the same cluster. Other works construct more comprehensive models based on the idea of clustering [3,22]. As in [22], the video is transformed into an undirected graph, and the summary is generated by partitioning this graph into clusters.…”
Section: Related Workmentioning
confidence: 99%
“…Besides, clustering is combined with other techniques to generate a summary, such as scene recognition in [14]. Recently, a co-clusters approach is proposed in [28], which summarizes several videos of the same topic simultaneously by identifying similar shots shared across these videos.…”
Section: Related Workmentioning
confidence: 99%
“…To describe the static and dynamic information of a video segment, we extract two types of features as the segment-level feature mapping [8, 19]: observation features extracted from a single frame, and interaction features extracted from two consecutive frames. Suppose the j th frame is described as a feature vector x j .…”
Section: Unsupervised Synchrony Discovery (Usd)mentioning
confidence: 99%
“…As a result, we represent a video segment X i = { x bi , ..., x ei } between the bith and the eith frames by normalizing the sum of the concatenation of the two features, resulting in a feature vector ϕboldXi=j=biei[ϕobs(boldxj);ϕint(boldxj)]. See [8,19] for details about the feature mapping.…”
Section: Unsupervised Synchrony Discovery (Usd)mentioning
confidence: 99%