ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9052966
|View full text |Cite
|
Sign up to set email alerts
|

Mspec-Net : Multi-Domain Speech Conversion Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 9 publications
(10 citation statements)
references
References 16 publications
0
10
0
Order By: Relevance
“…While all papers supported their results with at least one objective metric, only a few provided a subjective evaluation, namely: Lian et al (2019a), Parmar et al (2019), Patel et al (2021), Malaviya et al (2020), andPatel et al (2019). The lack of a subjective evaluation is justified in Niranjan et al (2020) since the VC was implemented in the context of ASR and having machine intelligibility in mind.…”
Section: Subjective Metricsmentioning
confidence: 99%
See 3 more Smart Citations
“…While all papers supported their results with at least one objective metric, only a few provided a subjective evaluation, namely: Lian et al (2019a), Parmar et al (2019), Patel et al (2021), Malaviya et al (2020), andPatel et al (2019). The lack of a subjective evaluation is justified in Niranjan et al (2020) since the VC was implemented in the context of ASR and having machine intelligibility in mind.…”
Section: Subjective Metricsmentioning
confidence: 99%
“…In Malaviya et al (2020), a multi-domain speech conversion system is proposed, capable of converting both from Non-Audible Murmur (NAM) and from whispered speech to normal speech, through three domain-specific AutoEncoders (AEs). These AEs are used to obtain an internal representation of features, which are known as latent representations.…”
Section: Mspec-netmentioning
confidence: 99%
See 2 more Smart Citations
“…Such as LSTM, MSpeC-Net, DiscoGAN, CycleGAN, etc. are proposed in the literature [2], [9], [12], [19]- [24]. Moreover, CycleGAN has shown state-of-the-art result for WHSP2SPCH conversion including F 0 prediction on parallel data, which relies on the availability of particular speaker's whisper, and normal speech [25].…”
Section: Introductionmentioning
confidence: 99%