Proceedings of the 3rd International on Multimodal Sentiment Analysis Workshop and Challenge 2022
DOI: 10.1145/3551876.3554809
|View full text |Cite
|
Sign up to set email alerts
|

Hybrid Multimodal Feature Extraction, Mining and Fusion for Sentiment Analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
15
0
2

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 28 publications
(17 citation statements)
references
References 22 publications
0
15
0
2
Order By: Relevance
“…As a fusion of both, Hybrid fusion combines the advantages of both early fusion and late fusion. Li et al (Li et al 2022) proposes a hybrid fusion method which leads them to the winner of MuSe-Reaction competition. In (Li et al 2022), audio features , facial expression features and paragraph-level text embeddings are fused at the feature level and then fed into the MMA module to extract complementary information from different modalities and calculate interactions between modalities.…”
Section: Multimodal Fusionmentioning
confidence: 99%
“…As a fusion of both, Hybrid fusion combines the advantages of both early fusion and late fusion. Li et al (Li et al 2022) proposes a hybrid fusion method which leads them to the winner of MuSe-Reaction competition. In (Li et al 2022), audio features , facial expression features and paragraph-level text embeddings are fused at the feature level and then fed into the MMA module to extract complementary information from different modalities and calculate interactions between modalities.…”
Section: Multimodal Fusionmentioning
confidence: 99%
“…Participants could utilise the text, audio, and video modality. Several systems were proposed to tackle this task, all of them employing the Transformer architecture [49,50,51].…”
Section: Multimodal Humour Recognitionmentioning
confidence: 99%
“…The TEMMA model proposed by Li [20] at Muse 2022 won the championship in this challenge. Resnet-18 feature and DeepSpectrum feature were used in [20].…”
Section: Emotional Reaction Intensitymentioning
confidence: 99%
“…The TEMMA model proposed by Li [20] at Muse 2022 won the championship in this challenge. Resnet-18 feature and DeepSpectrum feature were used in [20]. The researchers [20] use a Resnet-18 model to extract static visual features from the videos, and a DeepSpectrum [2] model to extract audio features from the videos.…”
Section: Emotional Reaction Intensitymentioning
confidence: 99%
See 1 more Smart Citation