2022
DOI: 10.1109/taffc.2020.2970418
|View full text |Cite
|
Sign up to set email alerts
|

Depression Level Prediction Using Deep Spatiotemporal Features and Multilayer Bi-LTSM

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
41
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 49 publications
(41 citation statements)
references
References 36 publications
0
41
0
Order By: Relevance
“…Most existing video-based automatic depression analysis approaches are single-stage methods, i.e., extracting depression feature from a single frame/thin slice or the entire video. In particular, the frame/thin slice-level methods attempted to model depression status based on individuals' facial appearance [13], [16], [17] (e.g., frame-level modelling) or short-term facial behaviours (thin slice-level modelling) [14], [15], [19], [36], [37], [38], [39]. The frame-level modelling approaches usually focus on learning the depressionrelated salient facial appearance information.…”
Section: Video-based Automatic Depression Analysismentioning
confidence: 99%
See 3 more Smart Citations
“…Most existing video-based automatic depression analysis approaches are single-stage methods, i.e., extracting depression feature from a single frame/thin slice or the entire video. In particular, the frame/thin slice-level methods attempted to model depression status based on individuals' facial appearance [13], [16], [17] (e.g., frame-level modelling) or short-term facial behaviours (thin slice-level modelling) [14], [15], [19], [36], [37], [38], [39]. The frame-level modelling approaches usually focus on learning the depressionrelated salient facial appearance information.…”
Section: Video-based Automatic Depression Analysismentioning
confidence: 99%
“…1, these methods fail to consider the important long-term facial behaviours/dynamics for depression recognition. Although some of the methods [15], [15], [16], [19] uses RNNs/LSTMs to model long-term temporal dependencies from the video, the CNNs of such methods are trained by pairing a frame/thin slice with the video-level label are problematic.…”
Section: Video-based Automatic Depression Analysismentioning
confidence: 99%
See 2 more Smart Citations
“…Similarly, VGG-LSTM and C3D-LSTM are employed in [11]. However, these approaches exhibit relatively weak performance regarding capturing facial dynamics [12]. More-over, on large-scale video data, deep learning-based methods experience computational efficiency issues.…”
Section: Introductionmentioning
confidence: 99%