2018
DOI: 10.1163/22134808-00002560
|View full text |Cite
|
Sign up to set email alerts
|

The Time Course of Audio-Visual Phoneme Identification: a High Temporal Resolution Study

Abstract: International audienc

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 58 publications
0
10
0
Order By: Relevance
“…In other words, when presented visually with a facial emotional expression, we need a shorter exposure to the stimulus compared to the matching vocal expression in order to reach the same efficient discrimination. In the field of word recognition, where the gating task has been more widely exploited, it has been shown that when discriminating a set of words that only differ by one phoneme, the performance can be better either in the visual or in the auditory domain depending on the saliency of the modality for each specific phoneme, and multisensory integration does not necessarily lead to a more successful discrimination (Sánchez-García et al 2018). Instead, in the context of emotion expressions' discrimination, our results robustly show that a multimodal context is always advantageous, and a discriminatory decision is reached earlier than in either unisensory condition.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…In other words, when presented visually with a facial emotional expression, we need a shorter exposure to the stimulus compared to the matching vocal expression in order to reach the same efficient discrimination. In the field of word recognition, where the gating task has been more widely exploited, it has been shown that when discriminating a set of words that only differ by one phoneme, the performance can be better either in the visual or in the auditory domain depending on the saliency of the modality for each specific phoneme, and multisensory integration does not necessarily lead to a more successful discrimination (Sánchez-García et al 2018). Instead, in the context of emotion expressions' discrimination, our results robustly show that a multimodal context is always advantageous, and a discriminatory decision is reached earlier than in either unisensory condition.…”
Section: Discussionmentioning
confidence: 99%
“…Indeed, emotion expression from the face and voice are intrinsically time-embedded and the accumulation of sensory evidence allowing for a reliable decision about which emotion is being displayed may vary across the senses and across emotions. The present study has therefore been designed with the scope of evaluating how observers accumulate informational evidence at different time points during the unfolding of dynamic visual, auditory and bimodal emotional signals, using a gating paradigm (Grosjean 1980;Jesse and Massaro 2010;Sánchez-García et al 2018). Building on existing cognitive models of speech perception (Marslen-Wilson and Welsh 1978;Marslen-Wilson 1987;Davis et al 2002), we assume that when a perceiver performs the extraction of emotional information from the face and/or voice of an interlocutor, discrete stored properties of each emotion matching the incoming expression are rapidly and partially activated, similarly to the activation of the "word initial cohort" when hearing incoming speech (Marslen-Wilson 1987).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…While the CIMS model is agnostic with respect to neuroanatomy, BOLD fMRI and modeling suggest that there are anatomical dissociations between brain areas responsible for sensory encoding and those responsible for causal inference judgments (Rohe and Noppeney, 2015;Cuppini et al, 2017) making it reasonable that individual differences in one computation are uncoupled from individual differences in the other. Findings that the McGurk effect shows different neural signatures than congruent audiovisual syllables (Erickson et al, 2014;Moris Fernandez et al, 2017;Sánchez-García et al, 2018) has been used as evidence that the McGurk effect is processed differently than everyday speech. The CIMS model clarifies that…”
Section: Relating the Mcgurk Effect To Other Speech Perception Tasksmentioning
confidence: 99%
“…The conceptual model explains the absence of multisensory benefit for voice-leading speech because of the lack of a perceptual head start provided by visual speech, suggesting a number of interesting experiments. Voice-leading speech could be transformed by experimentally manipulating auditory-visual asynchrony, advancing the visual portion of the recording and rendering it effectively "mouth-leading" (Magnotti et al, 2013;Sánchez-García et al, 2018).…”
Section: Model Predictions and Summarymentioning
confidence: 99%