1987
DOI: 10.3109/03005368709077786
|View full text |Cite
|
Sign up to set email alerts
|

Quantifying the contribution of vision to speech perception in noise

Abstract: The intelligibility of sentences presented in noise improves when the listener can view the talker's face. Our aims were to quantify this benefit, and to relate it to individual differences among subjects in lipreading ability and among sentences in lipreading difficulty. Auditory and audiovisual speech-reception thresholds (SRTs) were measured in 20 listeners with normal hearing. Sixty sentences, selected to range in the difficulty with which they could be lipread (with vision alone) from easy to hard, were p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

14
255
2

Year Published

2009
2009
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 370 publications
(271 citation statements)
references
References 12 publications
14
255
2
Order By: Relevance
“…Pretesting had established that this signal-to-noise ratio produced approximately 70% correct responses to auditory stimuli presented without visual speech. This point on the performance scale was chosen to provide appropriate conditions for revealing effects of display type on the influences exerted by congruent and incongruent visual speech on auditory speech perception (see, e.g., Jordan & Sergeant, 1998;MacLeod & Summerfield, 1987, 1990Middleweerd & Plomp, 1987;Sumby & Pollack, 1954). All sound measurements were made with a decibelometer at the location of each participant.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Pretesting had established that this signal-to-noise ratio produced approximately 70% correct responses to auditory stimuli presented without visual speech. This point on the performance scale was chosen to provide appropriate conditions for revealing effects of display type on the influences exerted by congruent and incongruent visual speech on auditory speech perception (see, e.g., Jordan & Sergeant, 1998;MacLeod & Summerfield, 1987, 1990Middleweerd & Plomp, 1987;Sumby & Pollack, 1954). All sound measurements were made with a decibelometer at the location of each participant.…”
Section: Methodsmentioning
confidence: 99%
“…The facial movements that accompany speech production (visual speech) are a powerful component of speech perception (e.g., Erber, 1969;Jordan, McCotter, & Thomas, 2000;MacLeod & Summerfield, 1987, 1990McGurk & MacDonald, 1976;Middleweerd & Plomp, 1987;Reisberg, McLean, & Goldfield, 1987;Rosenblum, Yakel, & Green, 2000;Sumby & Pollack, 1954;Thomas & Jordan, 2002). In particular, seeing the articulating face of a talker can improve auditory speech intelligibility substantially in quiet and noisy environments, and in the McGurk effect (after McGurk & MacDonald, 1976), can alter the perceived identity of speech sounds.…”
Section: Introductionmentioning
confidence: 99%
“…Finally, it should be noted that the speechreading performance scores obtained from our participants were rather low in comparison to most other studies using equivalent (BKB) stimulus materials. For instance, MacLeod & Summerfield (1987) found average performance ranged from around 20 to 50% of keywords, compared to the less than 10% obtained in the current study. This is a relatively common problem in tests of speechreading performance, with authors utilising various methods to improve performance; such as the inclusion of a low level of auditory speech (Landers & Davis, 2008).…”
Section: Discussionmentioning
confidence: 73%
“…'Mother baked a cake' and have been utilised in several studies of speechreading performance (e.g. MacLeod & Summerfield, 1987).…”
Section: Stimulimentioning
confidence: 99%
“…The benefit gained from the visual, facial cues has been quantitatively estimated to be equivalent to an increase of 8-10 dB in the signalto-noise ratio when speech sentences are presented in a noise background [9]. This observation suggests that, if the acoustic inputs to conventional speech recognition systems could be augmented by data about the visible speech gestures, an enhanced-performance, audio-visual recognition system should be possible.…”
Section: Introductionmentioning
confidence: 98%