2012 46th Annual Conference on Information Sciences and Systems (CISS) 2012
DOI: 10.1109/ciss.2012.6310945
|View full text |Cite
|
Sign up to set email alerts
|

A temporal saliency map for modeling auditory attention

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
18
0

Year Published

2013
2013
2022
2022

Publication Types

Select...
9
1

Relationship

3
7

Authors

Journals

citations
Cited by 28 publications
(19 citation statements)
references
References 13 publications
1
18
0
Order By: Relevance
“…One of the first models to address this problem computed a temporal salience map similarly to the model of Kayser et al ., but considered all of the features as evolving temporally, rather than as two-dimensional images [37]. The feature space was expanded to include perceptual properties of sound: loudness, pitch and timbre.…”
Section: Models Of Auditory Attentionmentioning
confidence: 99%
“…One of the first models to address this problem computed a temporal salience map similarly to the model of Kayser et al ., but considered all of the features as evolving temporally, rather than as two-dimensional images [37]. The feature space was expanded to include perceptual properties of sound: loudness, pitch and timbre.…”
Section: Models Of Auditory Attentionmentioning
confidence: 99%
“…Similar to visual saliency models (see section 3.2), auditory saliency models are built by abstracting features (intensity, frequency contrast, and temporal contrast) from the sound "intensity image, " which is a visual conversion of auditory time-frequency spectrograms and normalized to be an integrated saliency map (Kayser et al, 2005;Kalinli and Narayanan, 2007) (see Figure 2B). Considering that humans and other primate animals can process the pure auditory signals without any visual conversion, Kaya and Elhilali (2012) modify the auditory saliency model by directly extracting the multi-dimensional temporal auditory signal features (envelope, frequency, rate, bandwidth, and pitch) of the auditory scene as input. Their model relies on the selection of parameters to reduce error rates of the saliency determination by fewer features.…”
Section: Computational Models For the Human Cocktail Party Problem Somentioning
confidence: 99%
“…The nature of sound as a time-evolving entity cannot be captured by spatial processing. There have been attempts to remedy this problem by changes to the procedure of computing saliency after feature extraction, but the methodologies used are still adaptations from vision mechanisms (Kaya and Elhilali, 2012 ; Cottrell and Tsuchida, 2012 ). In this work, we discard the traditional framework of computing a spatial saliency map, and employ psychoacoustical experimentation and computational modeling to build a saliency extraction mechanism that broadly mimics processes that are hyphothesized to take place in the auditory pathway.…”
Section: Introductionmentioning
confidence: 99%