2018
DOI: 10.1121/1.5055562
|View full text |Cite
|
Sign up to set email alerts
|

A deep learning based segregation algorithm to increase speech intelligibility for hearing-impaired listeners in reverberant-noisy conditions

Abstract: Recently, deep learning based speech segregation has been shown to improve human speech intelligibility in noisy environments. However, one important factor not yet considered is room reverberation, which characterizes typical daily environments. The combination of reverberation and background noise can severely degrade speech intelligibility for hearing-impaired (HI) listeners. In the current study, a deep learning based time-frequency masking algorithm was proposed to address both room reverberation and back… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

2
24
0
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 37 publications
(27 citation statements)
references
References 38 publications
2
24
0
1
Order By: Relevance
“…Fortunately, deep learning algorithms based on timefrequency (T-F) masking have been successful at improving intelligibility under these varied conditions, and particularly for the population of greatest need-those with sensorineural hearing loss who wear hearing aids. Intelligibility improvements have been observed for backgrounds consisting of steady-state noise (Healy et al, 2013;Healy et al, 2014;Monaghan et al, 2017;Zhao et al, 2018), speech babble (Healy et al, 2013;Healy et al, 2014;Healy et al, 2015;Chen et al, 2016;Monaghan et al, 2017;Bentsen et al, 2018;Zhao et al, 2018), and cafeteria-noise recordings (Healy et al, 2015;Chen et al, 2016;Zhao et al, 2018).…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations
“…Fortunately, deep learning algorithms based on timefrequency (T-F) masking have been successful at improving intelligibility under these varied conditions, and particularly for the population of greatest need-those with sensorineural hearing loss who wear hearing aids. Intelligibility improvements have been observed for backgrounds consisting of steady-state noise (Healy et al, 2013;Healy et al, 2014;Monaghan et al, 2017;Zhao et al, 2018), speech babble (Healy et al, 2013;Healy et al, 2014;Healy et al, 2015;Chen et al, 2016;Monaghan et al, 2017;Bentsen et al, 2018;Zhao et al, 2018), and cafeteria-noise recordings (Healy et al, 2015;Chen et al, 2016;Zhao et al, 2018).…”
Section: Introductionmentioning
confidence: 99%
“…But the two types of distortion disrupt the acoustic speech signal in different ways, leading to different effects on speech perception (producing for example different patterns of vowel-perception errors in human listeners, N ab elek and Dagenais, 1986). Likely due to the considerable challenge associated with addressing these concurrent distortions, only recently reported is the first demonstration of intelligibility improvement resulting from a single-microphone (monaural) algorithm in reverberant-noisy conditions (Zhao et al, 2018). In this report, deep learning was used to estimate T-F masks for sentences corrupted by reverberation plus speech-shaped noise or reverberation plus multi-talker babble.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…The SE process consists of two parts: to enhance the intelligibility and quality of processed speech, and to reduce the noises in the background. Previous well-established algorithms have helped improve the SE in CI users [37], [38], [29], [39], [40], [41], [42], [43] but there are only few studies with a newly upgrading deep-learning-based algorithm. Traditional SE methods are based on identifying the difference between clean and noisy speech [44], [45], [46], [47], [48], [49].…”
mentioning
confidence: 99%