2019
DOI: 10.1038/s41598-019-44311-2
|View full text |Cite
|
Sign up to set email alerts
|

Combined predictive effects of sentential and visual constraints in early audiovisual speech processing

Abstract: In language comprehension, a variety of contextual cues act in unison to render upcoming words more or less predictable. As a sentence unfolds, we use prior context ( sentential constraints ) to predict what the next words might be. Additionally, in a conversation, we can predict upcoming sounds through observing the mouth movements of a speaker ( visual constraints ). In electrophysiological studies, effects of visual constraints have typically been observed early… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
2
0

Year Published

2020
2020
2025
2025

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 74 publications
2
2
0
Order By: Relevance
“…In agreement with prior studies (Solberg Økland et al, 2019;Hosaka et al, 2021), the N1 amplitudes elicited by the social audio recordings (human voices) in Experiment 2 were larger than those elicited by the non-social audio recordings (non-human sounds). Since the N1 component in the auditory modality is an index of early auditory processing (Zhao et al, 2016), and focused auditory attention could result in larger N1 amplitudes in response to audio recordings (Tumber et al, 2014), our findings suggest that social audio recordings capture more attention than non-social audio recordings.…”
Section: Methodssupporting
confidence: 91%
“…In agreement with prior studies (Solberg Økland et al, 2019;Hosaka et al, 2021), the N1 amplitudes elicited by the social audio recordings (human voices) in Experiment 2 were larger than those elicited by the non-social audio recordings (non-human sounds). Since the N1 component in the auditory modality is an index of early auditory processing (Zhao et al, 2016), and focused auditory attention could result in larger N1 amplitudes in response to audio recordings (Tumber et al, 2014), our findings suggest that social audio recordings capture more attention than non-social audio recordings.…”
Section: Methodssupporting
confidence: 91%
“…In agreement with prior studies [67,68], the N1 amplitudes elicited by the social audio recordings (human voices) in Experiment 2 were larger than those elicited by the nonsocial audio recordings (non-human sounds). Since the N1 component in the auditory modality is an index of early auditory processing [69], and focused auditory attention could result in larger N1 amplitudes in response to audio recordings [33], our findings suggest that social audio recordings capture more attention than nonsocial audio recordings.…”
Section: Methodssupporting
confidence: 91%
“…In contrast to iconic gestures, mouth movements are primarily useful in decoding the phonological information and listeners benefit from audiovisual speech because facial gestures can support predictions for upcoming words (Solberg Økland et al, 2019). This has been captured by the notion of visemes-that is, the shape(s) of the lips that correspond to a particular phoneme or group of phonemes (Fisher, 1968;Massaro & Cohen, 1995).…”
Section: Mouth Movementsmentioning
confidence: 99%
“…The majority of previous studies have only looked at the impact of one visual cue: iconic gestures or mouth movements, while the other cue was eliminated to achieve control. Thus, the face is cropped or covered in studies of gestures (e.g., Drijvers & Özyürek, 2017;Habets et al, 2010;Hirata & Kelly, 2010;Holle & Gunter, 2007;Holle et al, 2010), and the hands are not visible in studies of audiovisual speech (e.g., Ross et al, 2007;Solberg Økland et al, 2019;Tye-Murray et al, 2007). Only a handful of studies have investigated both gestures and mouth movements (Drijvers & Özyürek, 2017(Drijvers & Özyürek, , 2020Drijvers et al, 2019;Hirata & Kelly, 2010;Skipper et al, 2009;Zhang, Ding, et al, 2021a;Zhang, Frassinelli, et al, 2021b).…”
Section: Weighting the Multimodal Cuesmentioning
confidence: 99%