ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9053012
|View full text |Cite
|
Sign up to set email alerts
|

Gated Mechanism for Attention Based Multi Modal Sentiment Analysis

Abstract: Multimodal sentiment analysis has recently gained popularity because of its relevance to social media posts, customer service calls and video blogs. In this paper, we address three aspects of multimodal sentiment analysis; 1. Cross modal interaction learning, i.e. how multiple modalities contribute to the sentiment, 2. Learning long-term dependencies in multimodal interactions and 3. Fusion of unimodal and cross modal cues. Out of these three, we find that learning cross modal interactions is beneficial for th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 69 publications
(28 citation statements)
references
References 16 publications
0
28
0
Order By: Relevance
“…We compare the evaluation results of the model on CMU-MOSEI dataset with Graph-MFN [14], B2 + B4 w/multimodal fusion [16], Multilogue-Net [18], and TBJE [19]. e results of 2-classsentiment are shown in Table 1.…”
Section: E Results Of Cmu-mosei Datasetmentioning
confidence: 99%
See 1 more Smart Citation
“…We compare the evaluation results of the model on CMU-MOSEI dataset with Graph-MFN [14], B2 + B4 w/multimodal fusion [16], Multilogue-Net [18], and TBJE [19]. e results of 2-classsentiment are shown in Table 1.…”
Section: E Results Of Cmu-mosei Datasetmentioning
confidence: 99%
“…It utilizes self-attention to capture long term context and gating mechanism to selectively learn cross attended features [16].…”
Section: B2 + B4 W/multimodal Fusionmentioning
confidence: 99%
“…A gated mechanism could be considered as a special variant of attention mechanism, which also be employed for the cross-modal fusion. Kumar et al [ 46 ] proposed a conditional gated mechanism to modulate the information during mining inter-modal interaction.…”
Section: Related Workmentioning
confidence: 99%
“…A spoken interaction additionally requires conversational and channel understandability as highlighted in this work. While works have been carried out in understanding disfluency (Wang et al, 2020a;Lin and Wang, 2020) and turn-taking (Aldeneh et al, 2018;Hara et al, 2018), the authors narrowly aim at improving the task specific results by modelling acoustic cues (Aldeneh et al, 2018;Kumar and Vepa, 2020) or training with auxiliary tasks (Aldeneh et al, 2018;Hara et al, 2018;Wang et al, 2020a;Sundararaman et al, 2021). The effort in our work is orthogonal to what has been carried out in the past research.…”
Section: Related Workmentioning
confidence: 99%