Proceedings of the Tenth ACM International Conference on Multimedia - MULTIMEDIA '02 2002
DOI: 10.1145/641113.641116
|View full text |Cite
|
Sign up to set email alerts
|

A user attention model for video summarization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
237
0
3

Year Published

2005
2005
2024
2024

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 156 publications
(240 citation statements)
references
References 0 publications
0
237
0
3
Order By: Relevance
“…All visual attention algorithms generate a saliency map with a predicted pixel-level saliency. This saliency map is thresholded at k=1, 3,5,10,15,20,25 and 30 percent to obtain binary saliency maps. The percentage of human fixations contained within each binary map is the performance measure.…”
Section: Performancementioning
confidence: 99%
See 1 more Smart Citation
“…All visual attention algorithms generate a saliency map with a predicted pixel-level saliency. This saliency map is thresholded at k=1, 3,5,10,15,20,25 and 30 percent to obtain binary saliency maps. The percentage of human fixations contained within each binary map is the performance measure.…”
Section: Performancementioning
confidence: 99%
“…Therefore, understanding the manner in which humans process visual stimuli in a free viewing scenario has been an interesting problem in the scientific and engineering community. Several applications in computer vision (object recognition [1], visual tracking [2], text detection [3]), graphics (non-photo realistic rendering [4] ), multimedia (video summarization [5], video compression [6]) and robotics (robot localization [7]) can benefit from better understanding of human visual attention. A detailed overview of various saliency algorithms and its applications are presented in [8].…”
Section: Introductionmentioning
confidence: 99%
“…Going one step further toward human perception of saliency, Ma et al (2002) propose a framework for detecting the salient parts of a video based on user attention models. They use motion, face, and camera attention along with audio attention models (audio saliency and speech/music) as cues to capture salient information and identify the audio and video segments to compose the summary.…”
Section: Novelty Detection and Video Summarizationmentioning
confidence: 99%
“…The methods also tried to exploit the temporal component. Ma et al [4] defined a user attention model based on a motion vector field extracted from MPEG stream. This approach was used for video skimming.…”
Section: Introductionmentioning
confidence: 99%