Proceedings of the 2018 on Audio/Visual Emotion Challenge and Workshop 2018
DOI: 10.1145/3266302.3266311
|View full text |Cite
|
Sign up to set email alerts
|

Multi-modality Hierarchical Recall based on GBDTs for Bipolar Disorder Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
26
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 18 publications
(26 citation statements)
references
References 19 publications
0
26
0
Order By: Relevance
“…It is unlikely that performance will generalize as reported if studies did not evaluate performance on a representative held‐out test set (which are most studies in this review) and instead used some form of cross‐validation (which is the case in most reviewed studies), which is likely overfitting (see Figures and ). This limited generalizability and overfitting are observed for instance in the drop in performance from development to test set in submissions to the AVEC challenges . For results that used held‐out test sets, which are more likely to generalize if they are representative, scores range from close to chance to higher scores including Afshan et al (F1‐score = 0.95) which most likely benefited from having a large sample size (N depressed = 735, N controls = 953) and all participants being the same sex (female).…”
Section: Discussionmentioning
confidence: 99%
“…It is unlikely that performance will generalize as reported if studies did not evaluate performance on a representative held‐out test set (which are most studies in this review) and instead used some form of cross‐validation (which is the case in most reviewed studies), which is likely overfitting (see Figures and ). This limited generalizability and overfitting are observed for instance in the drop in performance from development to test set in submissions to the AVEC challenges . For results that used held‐out test sets, which are more likely to generalize if they are representative, scores range from close to chance to higher scores including Afshan et al (F1‐score = 0.95) which most likely benefited from having a large sample size (N depressed = 735, N controls = 953) and all participants being the same sex (female).…”
Section: Discussionmentioning
confidence: 99%
“…Iit is unlikely that performance will generalize as reported if studies did not evaluate performance on a representative held-out test set (which are most studies in this review) and instead used some form of cross-validation (which is the case in most reviewed studies), which is likely overfitting (see Figure 1 and Figure 5). This limited generalizability and overfitting are observed for instance in the drop in performance from development to test set in submissions to the AVEC challenges 8,9,60,63 . For results that used held-out test sets, which are more likely to generalize if they are representative, scores range from close to chance to higher scores including Afshan and colleagues (2018) 64 (F1-score = 0.95) which most likely benefited from having a large sample size (N depressed = 735, N controls = 953) and all participants being the same sex (female).…”
Section: Discussionmentioning
confidence: 99%
“…SVMs (A) [26] 55.0 50.0 GEWELMs (A) [30] 55.0 48.2 Multistream (A+V) [35] 78.3 40.7 IncepLSTM (A+V) [12] 65.1 -Hierarchical recall model(A+V) [33] 86.8 57.4 Multi-instance learning (A only) 61.6 57.4…”
Section: Uar[%] Dev T Estmentioning
confidence: 99%
“…The best result obtained by our proposed method is highly competitive with those obtained by state-of-the-art methods (Table 4). Except for [35] which also segmented the speech files, these methods fed features obtained from the full audio/video records into SVMs [26], Greedy Ensembles of Weighted Extreme Learning Machines (GEWELMs) [30], Inception LSTM (IncepLSTM) [12], or a hierarchical recall decision tree model [33]. Our multi-instance learning method performs better than all other audio-based methods, and better than, or comparable with, the methods using both speech and visual information.…”
Section: Uar[%] Dev T Estmentioning
confidence: 99%
See 1 more Smart Citation