2016
DOI: 10.1016/j.csl.2015.02.003
|View full text |Cite
|
Sign up to set email alerts
|

Data driven articulatory synthesis with deep neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
26
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 39 publications
(26 citation statements)
references
References 33 publications
0
26
0
Order By: Relevance
“…Current text-to-speech technologies have been able to produce speech with natural sounding voice for SSIs [19]. One of the current challenges of SSI development is silent speech recognition algorithms (without using audio data) [10,20] or mapping articulatory information to speech [21,22,23].…”
Section: Introductionmentioning
confidence: 99%
“…Current text-to-speech technologies have been able to produce speech with natural sounding voice for SSIs [19]. One of the current challenges of SSI development is silent speech recognition algorithms (without using audio data) [10,20] or mapping articulatory information to speech [21,22,23].…”
Section: Introductionmentioning
confidence: 99%
“…More details can be found in [50]. The choice for using a DNN-based mapping was motivated by previous results [35,51] showing that such mapping was more robust to noisy input than a state-of-the-art mapping based on Gaussian Mixture Model (GMM) such as the one proposed by Toda et al [34], and thus likely a better candidate for future BCI applications where input articulatory control signals will be inferred from noisy cortical signals.…”
Section: Methodsmentioning
confidence: 99%
“…In addition to the past and current source vectors, we also consider for the mapping ω future source vectors as this often improves the mapping accuracy at the expense of introducing a small delay in the conversion process [4,5,13]. Provided that this delay is less than 50 ms, we will be able to restore the articulatory-auditory feedback without causing disfluencies or mental stress on the speaker [14,15].…”
Section: Speech Synthesis From Articulator Movementmentioning
confidence: 99%
“…For comparison purposes, we also evaluate mappings using GMMs [4,25,26] and DNNs [13,27], which have been successfully applied by ourselves and other authors to model the articulatory-to-acoustic mapping. For a fair comparison, GMMs and DNNs with approximately the same number of parameters as the RNN architecture (˜1/2 million parameters) are employed.…”
Section: Model Trainingmentioning
confidence: 99%