This research developed a Multimodal Picture Word Task for assessing the influence of visual speech on phonological processing by100 children between 4 -14 yrs of age. We assessed how manipulation of seemingly to-be-ignored auditory (A) and audiovisual (AV) phonological distractors affected picture naming without participants consciously trying to respond to the manipulation. Results varied in complex ways as a function of age and type and modality of distractors. Results for congruent AV distractors yielded an inverted U-shaped function with a significant influence of visual speech in 4-yr-olds and 10-14-yr-olds, but not in 5-9-yr-olds. In concert with dynamic systems theory, we proposed that the temporary loss of sensitivity to visual speech was reflecting reorganization of relevant knowledge and processing sub-systems, particularly phonology. We speculated that reorganization may be associated with 1) formal literacy instruction and 2) developmental changes in multimodal processing and auditory perceptual, linguistic, and cognitive skills.
KeywordsPicture-Word Task; Audiovisual Speech Perception; U-Shaped Developmental Function; Phonological Processing; Picture-Word Interference; Picture Naming; Multimodal Speech Processing; Dynamic Systems Theory Speech communication by adults is naturally a multimodal event with auditory and visual speech integrated mandatorily. This basic property of mature speech perception is dramatically illustrated by McGurk effects (McGurk & McDonald, 1976). In the McGurk task, individuals hear a syllable whose onset has one place of articulation while seeing a talker simultaneously mouthing a syllable whose onset has a different place of articulation, e.g., auditory /ba/ and visual /ga/. Adults typically experience the illusion of perceiving /da/ or /ða/, a blend of the auditory and visual inputs. The McGurk illusion is consistent with the idea that auditory and visual speech interact prior to the classification of phonetic features, such as place of Address correspondence to: Susan Jerger, Ph.D., School of Behavioral and Brain Sciences, University of Texas at Dallas, P. O. Box 830688, GR4.1, Richardson, TX 75083-0688. Phone -972-883-2268; FAX -972-883-2491; sjerger@utdallas.edu. Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain. (Green, 1998). Integrating auditory and visual speech without conscious effort clearly has adaptive value. Seeing a talker's face facilitates listening in noisy soundscapes and in clear environments containing unfamiliar or complex content (Arnold & Hill, 2001;MacLeod & Summerfield, 1987;Massaro, 1998).
NIH Publ...