2019
DOI: 10.48550/arxiv.1909.11909
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Multichannel Speech Enhancement by Raw Waveform-mapping using Fully Convolutional Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 57 publications
0
2
0
Order By: Relevance
“…Although there are time-domain approaches that use multi-microphone modeling for speech enhancement and source separation [26]- [28], their effectiveness in environments with moderate to strong reverberation is not yet established [29]. In addition, our study tightly integrates multi-microphone complex spectral mapping with beamforming and post-filtering.…”
Section: Introductionmentioning
confidence: 99%
“…Although there are time-domain approaches that use multi-microphone modeling for speech enhancement and source separation [26]- [28], their effectiveness in environments with moderate to strong reverberation is not yet established [29]. In addition, our study tightly integrates multi-microphone complex spectral mapping with beamforming and post-filtering.…”
Section: Introductionmentioning
confidence: 99%
“…We note that such mismatch also exists in many other speech processing tasks, such as speech separation [26], where we observe that, by incorporating time-domain loss function [27], one can improve the output speech quality. More recently, deep learning approach to speech enhancement methods with time-domain raw waveform outputs [28,29] have also been investigated. However, we note that time-domain loss function has not been well explored in speech synthesis, which will be the focus of this paper.…”
Section: Introductionmentioning
confidence: 99%