ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019
DOI: 10.1109/icassp.2019.8683091
|View full text |Cite
|
Sign up to set email alerts
|

Phonetic Analysis of Dysarthric Speech Tempo and Applications to Robust Personalised Dysarthric Speech Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
62
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 56 publications
(63 citation statements)
references
References 23 publications
1
62
0
Order By: Relevance
“…Figure 1 shows how speaking rate and WER are correlated. We approximate speaking rate information by computing mean phoneme durations from forced alignments of the training data with the methodology of Xiong et al [22]. It can be seen that dysarthric speakers have the lowest speaking rates and also the highest WERs.…”
Section: -Wordmentioning
confidence: 99%
See 3 more Smart Citations
“…Figure 1 shows how speaking rate and WER are correlated. We approximate speaking rate information by computing mean phoneme durations from forced alignments of the training data with the methodology of Xiong et al [22]. It can be seen that dysarthric speakers have the lowest speaking rates and also the highest WERs.…”
Section: -Wordmentioning
confidence: 99%
“…typical speech [2,10,13,23]. Other works investigated transforming pathological speech to be more similar to typical speech, for example with speech enhancement methods [1] or by adjusting speech tempo [22]. Alternatively, Jiao et al [6], Xiong et al [22] have also employed data augmentation techniques to create additional, artifical dysarthric speech data.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…According to these assumptions, we propose a method to transfer the knowledge learned from these rich speech data sources to the problem of Japanese dysarthric speech recognition, for which training data is insufficient. Utilization of the speech data from physically unimpaired persons who speak the same language as the target speaker has been considered in some literatures [13]- [15]. However, these studies only considered the linguistic characteristics within the language.…”
Section: Depictsmentioning
confidence: 99%