Theory and implementation of acoustic virtual reality have matured and become a powerful tool for the simulation of entirely controllable virtual acoustic environments. Such virtual acoustic environments are relevant for various types of auditory experiments on subjects with normal hearing, facilitating flexible virtual scene generation and manipulation. When it comes to expanding the investigation group to subjects with hearing loss, choosing a reproduction system which offers a proper integration of hearing aids into the virtual acoustic scene is crucial. Current loudspeaker-based spatial audio reproduction systems rely on different techniques to synthesize a surrounding sound field, providing various possibilities for adaptation and extension to allow applications in the field of hearing aid-related research. Representing one option, the concept and implementation of an extended binaural real-time auralization system is presented here. This system is capable of generating complex virtual acoustic environments, including room acoustic simulations, which are reproduced as combined via loudspeakers and research hearing aids. An objective evaluation covers the investigation of different system components, a simulation benchmark analysis for assessing the processing performance, and end-to-end latency measurements.
Telephone conversation is ubiquitous within the office setting. Overhearing a telephone conversation-whereby only one of the two speakers is heard-is subjectively more annoying and objectively more distracting than overhearing a full conversation. The present study sought to determine whether this "halfalogue" effect is attributable to unexpected offsets and onsets within the background speech (acoustic unexpectedness) or to the tendency to predict the unheard part of the conversation (semantic [un]predictability), and whether these effects can be shielded against through top-down cognitive control. In Experiment 1, participants performed an office-related task in quiet or in the presence of halfalogue and dialogue background speech. Irrelevant speech was either meaningful or meaningless speech. The
Purpose Working memory capacity and language ability modulate speech reception; however, the respective roles of peripheral and cognitive processing are unclear. The contribution of individual differences in these abilities to utilization of spatial cues when separating speech from informational and energetic masking backgrounds in children has not yet been determined. Therefore, this study explored whether speech reception in children is modulated by environmental factors, such as the type of background noise and spatial configuration of target and noise sources, and individual differences in the cognitive and linguistic abilities of listeners. Method Speech reception thresholds were assessed in 39 children aged 5–7 years in simulated school listening environments. Speech reception thresholds of target sentences spoken by an adult male consisting of number and color combinations were measured using an adaptive procedure, with speech-shaped white noise and single-talker backgrounds that were either collocated (target and back-ground at 0°) or spatially separated (target at 0°, background noise at 90° to the right). Spatial release from masking was assessed alongside memory span and expressive language. Results and Conclusion Significant main effect results showed that speech reception thresholds were highest for informational maskers and collocated conditions. Significant interactions indicated that individual differences in memory span and language ability were related to spatial release from masking advantages. Specifically, individual differences in memory span and language were related to the utilization of spatial cues in separated conditions. Language differences were related to auditory stream segregation abilities in collocated conditions that lack helpful spatial cues, pointing to the utilization of language processes to make up for losses in spatial information.
This study considers whether bilingual children listening in a second language are among those on which higher processing and cognitive demands are placed when noise is present. Forty-four Swedish sequential bilingual 15 year-olds were given memory span and vocabulary assessments in their first and second language (Swedish and English). First and second language speech reception thresholds (SRTs) at 50% intelligibility for numbers and colors presented in noise were obtained using an adaptive procedure. The target sentences were presented in simulated, virtual classroom acoustics, masked by either 16-talker multi-talker babble noise (MTBN) or speech shaped noise (SSN), positioned either directly in front of the listener (collocated with the target speech), or spatially separated from the target speech by 90° to either side. Main effects in the Spatial and Noise factors indicated that intelligibility was 3.8 dB lower in collocated conditions and 2.9 dB lower in MTBN conditions. SRTs were unexpectedly higher by 0.9 dB in second language conditions. Memory span significantly predicted 17% of the variance in the second language SRTs, and 9% of the variance in first language SRTs, indicating the possibility that the SRT task places higher cognitive demands when listening to second language speech than when the target is in the listener's first language.
Auralization systems for auditory research should ideally be validated by perceptual experiments, as well as objective measures. This study employed perceptual tests to evaluate a recently proposed binaural real-time auralization system for hearing aid (HA) users. The dynamic localization of real sound sources was compared with that of virtualized ones, reproduced binaurally over headphones, loudspeakers with crosstalk cancellation (CTC) filters, research HAs, or combined via loudspeakers with CTC filters and research HAs under free-field conditions. System-inherent properties affecting localization cues were identified and their effects on overall horizontal localization, reversal rates, and angular error metrics were assessed. The general localization performance in combined reproduction was found to fall between what was measured for loudspeakers with CTC filters and research HAs alone. Reproduction via research HAs alone resulted in the highest reversal rates and angular errors. While combined reproduction helped decrease the reversal rates, no significant effect was observed on the angular error metrics. However, combined reproduction resulted in the same overall horizontal source localization performance as measured for real sound sources, while improving localization compared with reproduction over research HAs alone. Collectively, the results with respect to combined reproduction can be considered a performance indicator for future experiments involving HA users.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.