2019
DOI: 10.1109/tcsvt.2018.2860797
|View full text |Cite
|
Sign up to set email alerts
|

Action Parsing-Driven Video Summarization Based on Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
39
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 69 publications
(39 citation statements)
references
References 27 publications
0
39
0
Order By: Relevance
“…Finally, with respect to previously published works in IEEE TCSVT, our manuscript is most closely related to [17], [19], [25], [37] that suggest different deep-learning-based approaches for supervised video summarization. However, differently from them, our manuscript proposes a method that: i) learns summarization in a fully unsupervised manner, and ii) is the first to introduce the integration of a trainable AC model into a GAN to learn a policy for key-fragment selection and summarization.…”
Section: Relation Of the Proposed Methods With The Bibliographymentioning
confidence: 77%
See 1 more Smart Citation
“…Finally, with respect to previously published works in IEEE TCSVT, our manuscript is most closely related to [17], [19], [25], [37] that suggest different deep-learning-based approaches for supervised video summarization. However, differently from them, our manuscript proposes a method that: i) learns summarization in a fully unsupervised manner, and ii) is the first to introduce the integration of a trainable AC model into a GAN to learn a policy for key-fragment selection and summarization.…”
Section: Relation Of the Proposed Methods With The Bibliographymentioning
confidence: 77%
“…[14] uses video metadata for video categorization and to learn what is important in each category, and performs category-driven summarization by maximizing the relevance between the summary and the video's category. [15], [16], [17] similarly learn category-driven summarization in various ways, e.g., by using action classifiers. [18], [19] define a summary by maximizing its relevance with the video metadata, after projecting visual and textual data in a common latent space.…”
Section: A Supervised Video Summarizationmentioning
confidence: 99%
“…The comparison of the performance of the model on five state-of-the-art works by Gao et al [18], Muhammad et al [14], Muhammad et al [16], Muhammad et al [17], Liu et al [19] and Lei et al [21] based on f1-score are plotted in Figure 6(b). The results prove that the performance of the model surpasses most of the state-of-the-art works given the minimal number of classes trained and complexities involved.…”
Section: Number Of Key Frames Total Number Of Frames In the Video ð23þmentioning
confidence: 99%
“…This differs from prior work, which perform manually partitioning of the video into segments of the same length and do thereby not take the video content into account. Lei et al [26] propose another reinforcement learning-based summarization approach that also dynamically segments videos. The video is first segmented using a trained action classifier, so that each clip contains a single action, then a deep recurrent neural network is applied to select the most distinct frames for each clip.…”
Section: Video Summarization Using Reinforcement Learningmentioning
confidence: 99%
“…Reinforcement learning is used in this work to guide our summarization agent using a set of rewards that encode our underlying intuition of what qualities a successful summarization result should have. Deep reinforcement learning approaches [19][20][21] have been extensively used in a variety of computer vision tasks such as object segmentation [22], video captioning [23], action recognition [24], and also generic video summarization [25][26][27][28]. For example, Zhou and Qiao [25] develop a deep reinforcement learning-based summarization network with a diversity-representativeness reward to generate summaries, and achieve a good performance on generic video summarization.…”
Section: Introductionmentioning
confidence: 99%