2020
DOI: 10.1016/j.specom.2020.07.005
|View full text |Cite
|
Sign up to set email alerts
|

Parallel Representation Learning for the Classification of Pathological Speech: Studies on Parkinson’s Disease and Cleft Lip and Palate

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
25
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 30 publications
(25 citation statements)
references
References 37 publications
0
25
0
Order By: Relevance
“…Sustained vowels phonation is perhaps the most popular, also because it represents a very common task in different applications [ 24 ]. Other works focus on continuous speech recordings including sentences, read texts, and spontaneous speech, where clinically informative phenomena like prosody can be analyzed [ 19 , 37 , 48 ]. Few papers focused on the production of isolated words.…”
Section: Related Workmentioning
confidence: 99%
“…Sustained vowels phonation is perhaps the most popular, also because it represents a very common task in different applications [ 24 ]. Other works focus on continuous speech recordings including sentences, read texts, and spontaneous speech, where clinically informative phenomena like prosody can be analyzed [ 19 , 37 , 48 ]. Few papers focused on the production of isolated words.…”
Section: Related Workmentioning
confidence: 99%
“…We can differentiate two axes. On one side, the estimation of neurodegenerative pathologies that impact directly and typically the voice production of the subjects, such as pre-symptomatic Hungtington's disease [87% of accuracy reported in Rusz et al ( 3 )], Alzheimer's disease [80% of accuracy in Weiner et al ( 4 )], dysphonia [89% of accuracy in Tulics et al ( 5 )], or Parkinson's disease [84% of accuracy in Vasquez-Correa et al ( 6 )].…”
Section: Introductionmentioning
confidence: 99%
“…automatically learn high-level discriminative dysarthric representations [9][10][11][12][13][14]. Given the potential of deep learning-based approaches to characterize abstract but important acoustic cues beyond the realm of knowledge-based handcrafted features, in this paper we focus on deep learning-based approaches.…”
Section: Introductionmentioning
confidence: 99%
“…Mainstream deep learning-based dysarthric speech detection approaches rely on processing the magnitude spectrum (or features derived from the magnitude spectrum) of time-frequency representations such as the short-time Fourier transform (STFT) or continuous wavelet transform [10][11][12][13][14]. In [10], articulation impairments of patients suffering from dysarthria are modeled through a convolutional neural network (CNN) operating on the magnitude spectrum of the continuous wavelet transform.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation