2019
DOI: 10.1016/j.inffus.2018.06.003
|View full text |Cite
|
Sign up to set email alerts
|

Audio-visual emotion fusion (AVEF): A deep efficient weighted approach

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
54
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 127 publications
(54 citation statements)
references
References 20 publications
0
54
0
Order By: Relevance
“…However, transfer learning methods play a very limited role in this process. Common knowledge transfer in multi-modal methods include fine-tune well trained models to a specific type of signal (Vielzeuf et al, 2017;Yan et al, 2018;Huang et al, 2019;Ortega et al, 2019), or fine-tune different well-trained models to both speech and video signals (Ouyang et al, 2017;Zhang et al, 2017;Ma et al, 2019). Other usage of transfer learning for multi-modal methods includes leveraging the knowledge from one signal to another (e.g., video to speech) to reduce the potential bias (Athanasiadis et al, 2019).…”
Section: Multi-modal Transfer Learning For Emotion Recognitionmentioning
confidence: 99%
“…However, transfer learning methods play a very limited role in this process. Common knowledge transfer in multi-modal methods include fine-tune well trained models to a specific type of signal (Vielzeuf et al, 2017;Yan et al, 2018;Huang et al, 2019;Ortega et al, 2019), or fine-tune different well-trained models to both speech and video signals (Ouyang et al, 2017;Zhang et al, 2017;Ma et al, 2019). Other usage of transfer learning for multi-modal methods includes leveraging the knowledge from one signal to another (e.g., video to speech) to reduce the potential bias (Athanasiadis et al, 2019).…”
Section: Multi-modal Transfer Learning For Emotion Recognitionmentioning
confidence: 99%
“…However, bimodal emotion recognition can reach an accuracy of 86.85%, an increase of 5% compared with using a single modal of emotion recognition (Song et al 2015;Chuang and Wu 2004;Kessous et al 2010). Furthermore, previous studies have indicated that it is impossible to achieve satisfactory results by recognizing emotions based on a single model for either speech or facial expression (Ma et al 2019;Yang et al 2017). Accordingly, this study applied a bimodal emotion recognition system by using both facial expression recognition (55%) and speech recognition (45%).…”
Section: Methodsmentioning
confidence: 99%
“…Recently, a little work has investigated multiple modalities to recognize emotions [9][10][11][12]. Many studies have fused facial expression together with physiological signals [30]. Table 3 [30][31][32][33] presents some of the surveyed studies on the multi-modal emotion fusion, including the corpus, modalities of fusion, feature extractors, fusion approach, classifier, and classification accuracy.…”
Section: Literature Review On Multi-modal Emotion Recognitionmentioning
confidence: 99%
“…Many studies have fused facial expression together with physiological signals [30]. Table 3 [30][31][32][33] presents some of the surveyed studies on the multi-modal emotion fusion, including the corpus, modalities of fusion, feature extractors, fusion approach, classifier, and classification accuracy. As indicated in the literature and to the best of our knowledge, there are no works reported on fusing speech with EEG for recognizing emotions.…”
Section: Literature Review On Multi-modal Emotion Recognitionmentioning
confidence: 99%