2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC) 2018
DOI: 10.23919/apsipa.2018.8659591
|View full text |Cite
|
Sign up to set email alerts
|

Bayesian Multichannel Speech Enhancement with a Deep Speech Prior

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
37
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 35 publications
(37 citation statements)
references
References 15 publications
0
37
0
Order By: Relevance
“…When noisy environments are covered by training data used for DNN-based mask estimation, MNMF can be initialized by using the results of DNN-based mask estimation [2] to further refine the SCMs of speech and noise. It would be promising to use recently-proposed semi-supervised speech enhancement methods based on NMF or MNMF with a DNN-based prior on speech spectra [51]- [53].…”
Section: Discussionmentioning
confidence: 99%
“…When noisy environments are covered by training data used for DNN-based mask estimation, MNMF can be initialized by using the results of DNN-based mask estimation [2] to further refine the SCMs of speech and noise. It would be promising to use recently-proposed semi-supervised speech enhancement methods based on NMF or MNMF with a DNN-based prior on speech spectra [51]- [53].…”
Section: Discussionmentioning
confidence: 99%
“…[2,3,4,5,6]. Recently, deep generative speech models based on variational autoencoders (VAEs) [7] have been investigated for single-channel [8,9,10,11] and multichannel speech enhancement [12,13,14]. A pre-trained deep generative speech model is combined with a nonnegative matrix factorization (NMF) [15] noise model whose parameters are estimated at test time, from the observation of the noisy mixture signal only.…”
Section: Introductionmentioning
confidence: 99%
“…both single-channel [8][9][10] and multichannel scenarios [11][12][13][14][15]. The main idea of these studies is to use variational autoencoders (VAEs) [16] to replace the NMF generative model, thus benefiting from DNNs' representational power.…”
Section: Introductionmentioning
confidence: 99%
“…In [8][9][10][11][12], speech enhancement is performed using a pretrained VAE-based generative model of speech spectra combined with an unsupervised NMF model for the noise. Inference of the clean speech involves using the Metropolis-Hastings algorithm [19] to estimate an intractable distribution over the VAE's latent space.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation