Interspeech 2021 2021
DOI: 10.21437/interspeech.2021-753
|View full text |Cite
|
Sign up to set email alerts
|

Alzheimer Disease Recognition Using Speech-Based Embeddings From Pre-Trained Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
17
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(19 citation statements)
references
References 9 publications
0
17
0
Order By: Relevance
“…Text‐based features exhibit superior performance compared with acoustic‐based features [15, 16]. Specifically, in [22], the use of text‐based embedding vectors with pretrained BERT achieved the highest accuracy reported to date at 84.5%, whereas the accuracy using acoustic features remained at 74.6%.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Text‐based features exhibit superior performance compared with acoustic‐based features [15, 16]. Specifically, in [22], the use of text‐based embedding vectors with pretrained BERT achieved the highest accuracy reported to date at 84.5%, whereas the accuracy using acoustic features remained at 74.6%.…”
Section: Related Workmentioning
confidence: 99%
“…Among these studies, the 2020 ADReSS Challenge [11] used both speech signals and transcribed text, while the 2021 ADReSSo Challenge [12] focused on speech signals, and the 2023 ADReSS-M Challenge [13] considered multilingual speech. In this context, the ADReSSo challenges, which are of interest in this study, are primarily aimed at detecting dementia using acoustic features extracted from audio [15][16][17] or word-embedding representations [9,17,18,22] obtained from transcribed text acquired through ASR.…”
Section: Related Workmentioning
confidence: 99%
“…Pretrained speech representations for impaired speech is a relatively unexplored field. Positive results were obtained for Alzheimer's detection through speech with Wav2Vec [18,19]. However, limited studies have evaluated the effectiveness of speech representations for dysarthric speech recognition.…”
Section: Related Workmentioning
confidence: 99%
“…There is also information related to phones in the learned speech representations from the lower layers of the network. Therefore, the pre-trained models can be used to extract features in various speech-related tasks, such as in classification of stuttering and Alzheimer's disease, and in speaker and language identification [49,50,51,63,64].…”
Section: Feature Extraction Based On Pre-trained Modelsmentioning
confidence: 99%