2021 IEEE International Conference on Multimedia and Expo (ICME) 2021
DOI: 10.1109/icme51207.2021.9428142
|View full text |Cite
|
Sign up to set email alerts
|

Self-Attention Recurrent Summarization Network with Reinforcement Learning for Video Summarization Task

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 15 publications
(8 citation statements)
references
References 16 publications
0
8
0
Order By: Relevance
“…In another recent work, Yaliniz et al [30] used Independently Recurrent Neural Networks (IndRNNs) [18] to model the temporal dependence of video frames, and learned summarization by using rewards associated with the representativeness, diversity and uniformity (i.e., the temporal coherence) of the video summary. Finally, Phaphuangwittayakul et al [21] presented a variation of the network architecture from [34], which estimates the frames' importance by combining the created representations for the video frames at the output of a bi-directional RNN and a self-attention mechanism.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…In another recent work, Yaliniz et al [30] used Independently Recurrent Neural Networks (IndRNNs) [18] to model the temporal dependence of video frames, and learned summarization by using rewards associated with the representativeness, diversity and uniformity (i.e., the temporal coherence) of the video summary. Finally, Phaphuangwittayakul et al [21] presented a variation of the network architecture from [34], which estimates the frames' importance by combining the created representations for the video frames at the output of a bi-directional RNN and a self-attention mechanism.…”
Section: Related Workmentioning
confidence: 99%
“…where 𝜎 is the summary length regularization factor, a tunable hyper-parameter of our method (𝜎 was introduced in [19] and is used in several subsequent works, e.g., [3,17,21,34]). The computed training loss is then back-propagated to compute the gradients and update all the different parts of the architecture.…”
Section: Proposed Approachmentioning
confidence: 99%
See 3 more Smart Citations