2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR) 2018
DOI: 10.1109/mipr.2018.00043
|View full text |Cite
|
Sign up to set email alerts
|

Self-Attentive Feature-Level Fusion for Multimodal Emotion Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
21
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
7
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 47 publications
(21 citation statements)
references
References 18 publications
0
21
0
Order By: Relevance
“…Although this variation reduces the total parameter sizes of the network, it still does not benefit the model and gives a poorer performance to simple concatenation. Numerous other fusion methods such as tensor fusion (Zadeh et al, 2017), compact bilinear pooling (Gao et al, 2016), attention-based fusion (Poria et al, 2017;Hazarika et al, 2018), etc. are applicable, whose analyses, however, is not the focus of this paper.…”
Section: Resultsmentioning
confidence: 99%
“…Although this variation reduces the total parameter sizes of the network, it still does not benefit the model and gives a poorer performance to simple concatenation. Numerous other fusion methods such as tensor fusion (Zadeh et al, 2017), compact bilinear pooling (Gao et al, 2016), attention-based fusion (Poria et al, 2017;Hazarika et al, 2018), etc. are applicable, whose analyses, however, is not the focus of this paper.…”
Section: Resultsmentioning
confidence: 99%
“…Using deep learning to classify facial expressions is usually learning how to use strong supervision methods [15] Barros et al proposed a network model based on the topological structure of VGG-16 for the formalization of the Facial Channel neural network for Facial Expression Recognition (FER) [24]. Koujan et al proposed a CNN that recognized human emotions from a single face image [25].…”
Section: Related Workmentioning
confidence: 99%
“…MED means to perform emotion detection from a multimodal learning perspective [3], [26]. A large amount of current research, such as [27]- [31], have been dependent on multimodal techniques for emotion detection. These exhibit good performances in emotion detection systems [32]- [34], thus promoting the use of multimodality.…”
Section: Related Workmentioning
confidence: 99%