2016
DOI: 10.1016/j.neucom.2015.05.126
|View full text |Cite
|
Sign up to set email alerts
|

Divide-and-conquer based summarization framework for extracting affective video content

Abstract: Recent advances in multimedia technology have led to tremendous increases in the available volume of video data, thereby creating a major requirement for efficient systems to manage such huge data volumes. Video summarization is one of the key techniques for accessing and managing large video libraries. Video summarization can be used to extract the affective contents of a video sequence to generate a concise representation of its content. Human attention models are an efficient means of affective content extr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
19
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 39 publications
(19 citation statements)
references
References 46 publications
0
19
0
Order By: Relevance
“…Each technique summarizes the video using a specific feature, such as trajectories, moving objects, abnormal detection, and many others. These categories of techniques can be classified into two general categories, scene-based (i.e., static [1-7, 9, 17, 21, 22], dynamic [8,12,13,16,[18][19][20] and content-based approaches, and the content-based approaches can be further decomposed into three types related to the content of the video including motion-based [10][11][12][13][14][15][16]20], action-based [21,22] and event-based [11,15,[17][18][19], as shown in Fig. 1.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Each technique summarizes the video using a specific feature, such as trajectories, moving objects, abnormal detection, and many others. These categories of techniques can be classified into two general categories, scene-based (i.e., static [1-7, 9, 17, 21, 22], dynamic [8,12,13,16,[18][19][20] and content-based approaches, and the content-based approaches can be further decomposed into three types related to the content of the video including motion-based [10][11][12][13][14][15][16]20], action-based [21,22] and event-based [11,15,[17][18][19], as shown in Fig. 1.…”
Section: Related Workmentioning
confidence: 99%
“…Called HSUMM, the proposed approach adopts a hierarchical clustering method to generate a weight map from the frame similarity graph in which the clusters can easily be inferred. In the same context and to generate an efficient summarization, the authors in [8] proposed a divide-and-conquer-based framework. In this work, the original video data is divided into shots, where an attention model is computed from each shot in parallel.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The semantic gap refers to the gap between the actual semantic topic that a viewer infers from watching a video and the topic inferred from video summarization, which is constructed using the extraction algorithms. To bridge the semantic gap, visual‐attention‐model‐based summarization schemes have been proposed to extract frames as keyframes if they are visually important for humans, as determined using visual attention models (Ejaz, Mehmood, & Baik, ; Mehmood, Sajjad, Rho, & Baik, ). Mehmood et al () proposed a human‐attention model that combines both multimedia content and the viewer's neuronal responses for video summarization.…”
Section: Introductionmentioning
confidence: 99%
“…To bridge the semantic gap, visual‐attention‐model‐based summarization schemes have been proposed to extract frames as keyframes if they are visually important for humans, as determined using visual attention models (Ejaz, Mehmood, & Baik, ; Mehmood, Sajjad, Rho, & Baik, ). Mehmood et al () proposed a human‐attention model that combines both multimedia content and the viewer's neuronal responses for video summarization. In their model, neuronal attention is computed using beta‐band electroencephalography (EEG) frequencies of neuronal responses.…”
Section: Introductionmentioning
confidence: 99%