Proceedings of the 6th International Workshop on Audio/Visual Emotion Challenge 2016
DOI: 10.1145/2988257.2988261
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal and Multiresolution Depression Detection from Speech and Facial Landmark Features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
42
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 100 publications
(52 citation statements)
references
References 38 publications
0
42
0
Order By: Relevance
“…The VL-Formants also outperformed the i-vector paradigm depression F1-scores on the development and test sets (0.57 and 0.48, as presented in [28]). Impressively, they match performance with the DepAudioNet system presented in [29].…”
Section: Resultsmentioning
confidence: 81%
See 1 more Smart Citation
“…The VL-Formants also outperformed the i-vector paradigm depression F1-scores on the development and test sets (0.57 and 0.48, as presented in [28]). Impressively, they match performance with the DepAudioNet system presented in [29].…”
Section: Resultsmentioning
confidence: 81%
“…These include: VTC features [27]; the i-vector paradigm [28]; and a deep neural network which combined both convolutional and Long Short Term Memory (LSTM) layers [29].…”
Section: Depression Corpusmentioning
confidence: 99%
“…The second approach to modeling depression attempts to exploit global and/or time varying statistics, independent of the question that prompted the response. Williamson et al utilized correlations of formants and spectral information across different time scales [11], Syed et al developed audio and video features to capture temporal variations [12], while Pampouchidou et al and Nasir et al fused low and high-level features [13,14]. Utilizing emerging techniques, Ma et al used audio to model depression by allowing deep neural networks to learn such associations rather than perform feature engineering [15].…”
Section: Introductionmentioning
confidence: 99%
“…Their strengths lie in their ability to represent information through non-linear transforms, at varying spatial and temporal resolution, and from multiple modalities [17,18]. While work in the domain of detecting depression has looked at fusing features from multiple modalities together [9,13,14,19], and utilizing neural networks to model single sequences [10,15], there remains to explore the sequence modeling of depression that utilizes deep learning approaches.…”
Section: Introductionmentioning
confidence: 99%
“…These challenges hint at the potential benefits of involving automatic behavior annotation, in which data-driven machine learning techniques are employed to automatically extract behavioral information directly from data, rather than relying on time-consuming and expensive annotations from human experts. Such behavior analysis work has been shown to be effective at identifying behaviors during interactions in domains such as couple therapy [7,8,9], depression [10,11,12] and suicide risk assessment [13,14,15,16]. However, due to potential domain mismatch, obtaining accurate performance in one domain by utilizing well-trained behavior analysis systems from a different domain is not straightforward.…”
Section: Introductionmentioning
confidence: 99%