2016
DOI: 10.1109/lsp.2015.2509480
|View full text |Cite
|
Sign up to set email alerts
|

Supervised Monaural Speech Enhancement Using Complementary Joint Sparse Representations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
7
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 21 publications
(7 citation statements)
references
References 25 publications
0
7
0
Order By: Relevance
“…OVL is a linear combination of three objective measures including PESQ, log likelihood ratio (LLR) [21] and weighted-slope spectral (WSS) distance [22]. Moreover, we compare all the experimental results of our proposed method with five state-of-the-art speech enhancement methods involving GDL [6], CJSR [9], RPCA [11], LSLD [13] and CLSMD [10].…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…OVL is a linear combination of three objective measures including PESQ, log likelihood ratio (LLR) [21] and weighted-slope spectral (WSS) distance [22]. Moreover, we compare all the experimental results of our proposed method with five state-of-the-art speech enhancement methods involving GDL [6], CJSR [9], RPCA [11], LSLD [13] and CLSMD [10].…”
Section: Resultsmentioning
confidence: 99%
“…Different from the GDL method, the complementary joint sparse representation (CJSR) method [9] utilizes the noisy speech magnitude spectra in the learning stage together with the clean speech magnitude spectra and the noise magnitude spectra, respectively, to form two different training datasets to train two distinct mixture dictionaries. The trained mixture dictionaries offer latent mappings from the noisy speech to the clean speech and noise, respectively, which enable the sparse coding algorithm in the enhancement stage to improve the accuracy of both speech and noise estimations in the frequency domain.…”
Section: Introductionmentioning
confidence: 99%
“…Then, the clean speech PSD estimate is found based on multiplication of the sparse code with the dictionary. Luo et al [17] proposed a complementary joint sparse representation, where two mixture dictionaries to model "mixture and speech" and "mixture and noise" are added to the Generative Dictionary Learning (GDL) problem formulation, and sparse codes of clean speech are forced to represent the noisy mixture on the mixture and clean speech sub-dictionary, while the sparse codes for the noise are forced to represent the noisy mixture on the mixture and noise sub-dictionary. Though this joint sparse representation alleviates, to some extent, the problem of source confusion, it is characterized by high complexity due to the need of learning four subdictionaries instead of two.…”
Section: Introductionmentioning
confidence: 99%
“…Then the target type was determined by minimizing the reconstruction error of every type of training sample. Sparse representation has been used extensively in various recognition tasks, such as facial recognition, speech recognition, and hyperspectral image classification [ 10 , 11 , 12 ]. Sparse representation based recognition has gained in-depth development in accordance to different requirements of image recognition.…”
Section: Introductionmentioning
confidence: 99%