1985
DOI: 10.1121/1.392336
|View full text |Cite
|
Sign up to set email alerts
|

Intermodal timing relations and audio-visual speech recognition by normal-hearing adults

Abstract: Audio-visual identification of sentences was measured as a function of audio delay in untrained observers with normal hearing; the soundtrack was replaced by rectangular pulses originally synchronized to the closing of the talker's vocal folds and then subjected to delay. When the soundtrack was delayed by 160 ms, identification scores were no better than when no acoustical information at all was provided. Delays of up to 80 ms had little effect on group-mean performance, but a separate analysis of a subgroup … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

14
127
1

Year Published

1993
1993
2020
2020

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 165 publications
(142 citation statements)
references
References 0 publications
14
127
1
Order By: Relevance
“…Like previous works have shown [21,26,42], asynchrony can be detected at very short thresholds, likely depending on the nature of the audiovisual event. Our observations indicate that such caution would involve the avoidance of video stream delays entirely, since the windows of temporal integration for two of three content types do not extend to audio lead asynchrony, irrespective of distortion.…”
Section: Resultsmentioning
confidence: 97%
See 2 more Smart Citations
“…Like previous works have shown [21,26,42], asynchrony can be detected at very short thresholds, likely depending on the nature of the audiovisual event. Our observations indicate that such caution would involve the avoidance of video stream delays entirely, since the windows of temporal integration for two of three content types do not extend to audio lead asynchrony, irrespective of distortion.…”
Section: Resultsmentioning
confidence: 97%
“…While some experiments implement a temporal order judgement task, asking participants to determine which signal precedes the other [39], others rely on a simultaneity judgement task that requires assessments on perceived synchrony [5]. Still others ask for the detection of gradually introduced asynchrony [8], or discrimination between presentations with different temporal offsets [26]. Figure 1 provides a summary of previously published thresholds corresponding to perceived synchrony for different event types, but also established using different measures.…”
Section: Temporal Integration and Quality Distortionmentioning
confidence: 99%
See 1 more Smart Citation
“…Thus, the video token of the original tape and an edited audio token were recorded simultaneously onto a second tape, resulting in a new synchronous audiovisual token. The lag time for dubbing was found to be no greater than 9.4 msec, well below the 80-msec range required for observers to detect an audiovisual asynchrony (McGrath & Summerfield, 1985).…”
Section: Methodsmentioning
confidence: 99%
“…In the McGurk effect, a lipread speech token affects the phonetic content of a speech sound that is heard (McGurk & MacDonald, 1976). The effect was biggest when auditory vowels were synchronized with the original mouth movements (McGrath & Summerfield, 1985), but the effect survives, even if audition lagged vision by 180 msec (see also Soto-Faraco & Alsius, 2007; these studies show that participants can still perceive a McGurk effect when they can quite reliably perform TOJs at intervals above the JND). So, speech-sound identification can be influenced by lipread speech even if the two are perceived as being out of sync.…”
Section: The Window Of Temporal Integrationmentioning
confidence: 98%