For analysing the processing of speech by a hearing instrument, a standard test signal is necessary which allows for reproducible measurement conditions, and which features as many of the most relevant properties of natural speech as possible, e.g. the average speech spectrum, the modulation spectrum, the variation of the fundamental frequency together with its appropriate harmonics, and the comodulation in different frequency bands. Existing artificial signals do not adequately fulfill these requirements. Moreover, recordings from natural speakers represent only one language and are therefore not internationally acceptable. For this reason, an International Speech Test Signal (ISTS) was developed. It is based on natural recordings but is largely non-intelligible because of segmentation and remixing. When using the signal for hearing aid measurements, the gain of a device can be described at different percentiles of the speech level distribution. The primary intention is to include this test signal with a new measurement method for a new hearing aid standard (IEC 60118-15).
This study compared the combined effect of noise and reverberation on listening effort and speech intelligibility to predictions of the speech transmission index (STI). Listening effort was measured in normal-hearing subjects using a scaling procedure. Speech intelligibility scores were measured in the same subjects and conditions: (a) Speech-shaped noise as the only interfering factor, (b) + (c) fixed signal-to-noise ratios (SNRs) of 0 or 7 dB and reverberation as detrimental factors, and (d) reverberation as the only detrimental factor. In each condition, SNR and reverberation were combined to produce STI values of 0.17, 0.30, 0.43, 0.57, and 0.70, respectively. Listening effort always decreased with increasing STI, thus enabling a rough prediction, but a significant bias was observed indicating that listening effort was lower in reverberation only than in noise only at the same STI for one type of impulse responses. Accordingly, speech intelligibility increased with increasing STI and was significantly better in reverberation only than in noise only at the same STI. Further analyses showed that the broadband reverberation time is not always a good estimate of speech degradation in reverberation and that different speech materials may differ in their robustness toward detrimental effects of reverberation.
An adaptive procedure for controlling the signal-to-noise ratio (SNR) when rating the subjectively perceived listening effort (Adaptive Categorical Listening Effort Scaling) is described. For this, the listening effort is rated on a categorical scale with 14 steps after the presentation of three sentences in a background masker. In a first phase of the procedure, the individual SNR range for ratings from "no effort" to "extreme effort" is estimated. In the following phases, stimuli with randomly selected SNRs within this range are presented. One or two linear regression lines are fitted to the data describing subjective listening effort as a function of SNR. The results of the adaptive procedure are independent of the initial SNR. Although a static procedure using fixed, predefined SNRs produced similar results, the adaptive procedure avoided lengthy pretests for suitable SNRs and limited possible bias in the rating procedures. The adaptive procedure resolves individual differences, as well as differences between maskers. Inter-individual standard deviations are about three times as large as intra-individual standard deviations and the intra-class correlation coefficient for test-retest reliability is, on average, 0.9.
Sensorneural hearing-impaired listeners suffer severely from deterioration in the processing and internal representation of acoustic signals. In order to understand this deterioration in detail, a numerical perception model was developed which is based on current functional models of the signal processing in the auditory system. To test this model, the individual's speech intelligibility in quiet and in noise was predicted. The primary input parameter of the model is the precisely measured audiogram of each listener. In a refined version of the model, additional input parameters are derived from predicting the individual's temporal forward masking and notched-noise measurements with the same model assumptions. The predictions of the perception model were compared with those of the articulation index (AI) and the speech transmission index (STI). The accuracy of prediction with the perception model is in the same range as with the AI and the STI. The model does not require a calibration function and has the advantage of a greater flexibility in including different processing deficits associated with hearing impairment. However, it requires more time for computation. For the hearing-impaired listeners examined so far the individually measured psychoacoustical parameters have only a secondary effect on the prediction as compared to the audiogram. Nevertheless, the underlying model is a first step toward a quantitative understanding of speech intelligibility and helps to distinguish between the influence of the "attenuation" and the "distortion" component of the hearing loss.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.