2019
DOI: 10.1186/s13634-019-0611-y
|View full text |Cite
|
Sign up to set email alerts
|

A bottom-up summarization algorithm for videos in the wild

Abstract: Video summarization aims to provide a compact video representation while preserving the essential activities of the original video. Most existing video summarization approaches relay on identifying important frames and optimizing target energy by a global optimum solution. But global optimum may fail to express continuous action or realistically validate how human beings perceive a story. In this paper, we present a bottom-up approach named clip growing for video summarization, which allows users to customize … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(8 citation statements)
references
References 31 publications
0
8
0
Order By: Relevance
“…These works require the use of fixed cameras and suffer from immobility, significantly restricting the range of potential applications. The requirement of using obtrusive and non-wearable sensors to extract a user's emotions has encouraged most research to focus on direct affective content utilizing video frames [12][13][14][15], video segments [16][17][18][19], audio and text features [14]. The present results remained poor on average due to the semantic gap problem [10] of detected objects or events.…”
Section: Introductionmentioning
confidence: 75%
See 1 more Smart Citation
“…These works require the use of fixed cameras and suffer from immobility, significantly restricting the range of potential applications. The requirement of using obtrusive and non-wearable sensors to extract a user's emotions has encouraged most research to focus on direct affective content utilizing video frames [12][13][14][15], video segments [16][17][18][19], audio and text features [14]. The present results remained poor on average due to the semantic gap problem [10] of detected objects or events.…”
Section: Introductionmentioning
confidence: 75%
“…In contrast to previous works, where video highlight/summaries are generated with a single score function representing frame-level importance [2,10,[12][13][14][15]18,19], we trained two independent classifiers on highlight start and end times. This divide-and-conquer approach simplifies the task for each classifier by specializing in only one classification ( Figure 12).…”
Section: Machine Learning Approachmentioning
confidence: 99%
“…when comparing various models, CNN using only TDM characteristics is the superior option. It is claimed in Pan et al [2] that energy entropy-based summarization can be used to further refine this complicated model. Assigning an entropy block to each paired frame, we may calculate their energy using the following formulas:…”
Section: Literature Reviewmentioning
confidence: 99%
“…Video outline techniques ordinarily recognize frames that have higher development esteems, and thusly moving article recognition can be a component vector for this reason. The work in [1] utilizes Moving Object Detection and Image Similarity with the end goal of KFE. It utilizes ViBe calculation and breakers between frame distinction strategy by isolating the first video into a few portions that contain the moving item.…”
Section: Previous Work On Video Summarization Modelsmentioning
confidence: 99%