Though necessary, protective mask wearing in response to the COVID-19 pandemic presents communication challenges. The present study examines how signal degradation and loss of visual information due to masks affects intelligibility and memory for native and non-native speech. We also test whether clear speech can alleviate perceptual difficulty for masked speech. One native and one non-native speaker of English recorded video clips in conversational speech without a mask and conversational and clear speech with a mask. Native English listeners watched video clips presented in quiet or mixed with competing speech. The results showed that word recognition and recall of speech produced with a mask can be as accurate as without a mask in optimal listening conditions. Masks affected non-native speech processing at easier noise levels than native speech. Clear speech with a mask significantly improved accuracy in all listening conditions. Speaking clearly, reducing noise, and using surgical masks as well as good signal amplification can help compensate for the loss of intelligibility due to background noise, lack of visual cues, physical distancing, or non-native speech. The findings have implications for communication in classrooms and hospitals where listeners interact with teachers and healthcare providers, oftentimes non-native speakers, through their protective barriers.
Though necessary, protective mask wearing in response to the Covid-19 pandemic presents communication challenges. Speech produced through these masks is quieter, muffled, and lacks important visual cues for listeners. The present study examines how such loss of visual information and signal degradation affects intelligibility and auditory memory of non-native speech. Additionally, we examined whether clear speaking style can alleviate some perceptual difficulty for masked speech. One non-native English speaker recorded short, educationally relevant video clips in three conditions: casual speech without a mask, and casual and clear speech with a mask. In an online study, native English listeners were presented with 15 video clips in each of the three mask / style conditions and transcribed what they heard. Following each condition, listeners answered questions about the content of the clips. Word recognition and memory accuracy will be analyzed. Detailed acoustic analyses will be conducted to assess how face masks and speaking style affect speech production. The results will allow us to quantify speech communication challenges arising from the widespread use of masks. The findings will have implications for communication in classrooms and hospitals where listeners have to understand teachers and healthcare providers, often time non-native speakers, through their protective barriers.
The present study examined the effect of intelligibility-enhancing clear speech on listeners' recall. Native (n = 57) and non-native (n = 31) English listeners heard meaningful sentences produced in clear and conversational speech, and then completed a cued-recall task. Results showed that listeners recalled more words from clearly produced sentences. Sentence-level analysis revealed that listening to clear speech increased the odds of recalling whole sentences and decreased the odds of erroneous and omitted responses. This study showed that the clear speech benefit extends beyond word- and sentence-level recognition memory to include deeper linguistic encoding at the level of syntactic and semantic information.
The goal of the study was to examine whether enhancing the clarity of the speech signal through conversational-to-clear speech modifications improves sentence recognition memory for native and non-native listeners, and if so, whether this effect would hold when the stimuli in the test phase are presented in orthographic instead of auditory form (cross-modal presentation). Sixty listeners (30 native and 30 non-native English) participated in a within-modal (i.e., audio-audio) sentence recognition memory task (Experiment I). Sixty different individuals (30 native and 30 non-native English) participated in a cross-modal (i.e., audio-textual) sentence recognition memory task (Experiment II). The results showed that listener-oriented clear speech enhanced sentence recognition memory for both listener groups regardless of whether the acoustic signal was present during the test phase (Experiment I) or absent (Experiment II). Compared to native listeners, non-native listeners had longer reaction times in the within-modal task and were overall less accurate in the cross-modal task. The results showed that more cognitive resources remained available for storing information in memory during processing of easier-to-understand clearly produced sentences. Furthermore, non-native listeners benefited from signal clarity in sentence recognition memory despite processing speech signals in a cognitively more demanding second language.
Previous research has shown that native and non-native listeners’ recognition memory is higher for sentences previously heard in clear speech (CS) than in conversational speech (Keerstock and Smiljanic, 2018). The present study investigated whether speaking clearly also enhances talkers’ sentence recognition memory. The production effect (MacLeod et al., 2010) revealed superior retention of material read aloud relative to material read silently during an encoding phase. Vocal production that included an additional distinct element, such as speaking loudly, produced even greater memory benefits than reading aloud (Quinlan and Taylor, 2013). Production of the exaggerated articulatory and acoustic cues in CS could thus lead to enhanced sentence recognition memory relative to conversational style. Native and non-native English speakers produced alternating blocks of 10 conversational and CS sentences for a total of 60 unique sentences. In the test phase, they identified 120 sentences as old (from exposure) or new (distractors). Unexpectedly, preliminary results show lower sentence recognition memory for sentences produced in CS than in conversational speech for both groups of talkers. The results suggest that producing CS, unlike perceiving it, interferes with recognition memory. Allocating cognitive resources to producing hyper-articulated speech may limit their availability for storing information in memory.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.