2022
DOI: 10.3389/fnagi.2022.830943
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Deep Learning Models for Detecting Dementia From Speech and Transcripts

Abstract: Alzheimer's dementia (AD) entails negative psychological, social, and economic consequences not only for the patients but also for their families, relatives, and society in general. Despite the significance of this phenomenon and the importance for an early diagnosis, there are still limitations. Specifically, the main limitation is pertinent to the way the modalities of speech and transcripts are combined in a single neural network. Existing research works add/concatenate the image and text representations, e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 22 publications
(9 citation statements)
references
References 35 publications
0
9
0
Order By: Relevance
“…Ilias et al [ 149 ] combined a BERT with a ViT and Co-attention to successfully fulfill an AD-nonAD classification task.…”
Section: Resultsmentioning
confidence: 99%
“…Ilias et al [ 149 ] combined a BERT with a ViT and Co-attention to successfully fulfill an AD-nonAD classification task.…”
Section: Resultsmentioning
confidence: 99%
“… 33 Prior research indicates that approaches using acoustic speech input are also sensitive to dementia, albeit usually to a lesser extent than approaches using textual speech input. 43 Combining these different information modalities in speech could help to further augment classification sensitivity, 11 , 44 although this improvement is not always shown in practice. 45 Potential directions for future research in the AMYPRED sample include linguistic, temporal, and acoustic aspects of speech.…”
Section: Discussionmentioning
confidence: 99%
“…However, speech is highly multidimensional, incorporating not only linguistic but also acoustic and temporal features, which could confer additional sensitivity to cognitive impairment and amyloid positivity 33 . Prior research indicates that approaches using acoustic speech input are also sensitive to dementia, albeit usually to a lesser extent than approaches using textual speech input 43 . Combining these different information modalities in speech could help to further augment classification sensitivity, 11,44 although this improvement is not always shown in practice 45 .…”
Section: Discussionmentioning
confidence: 99%
“…They also performed occlusion analysis to reveal the most significant brain areas for the dementia subtypes classification task. Additionally, researchers also used cognitive data [56], speech features [57,58], and fundus photographs data [59] in deep learning models for dementia diagnosis.…”
Section: Deep Learning With Multimodal Data In Dementiamentioning
confidence: 99%