2022 IEEE International Conference on Multimedia and Expo (ICME) 2022
DOI: 10.1109/icme52920.2022.9860014
|View full text |Cite
|
Sign up to set email alerts
|

Utilizing BERT Intermediate Layers for Multimodal Sentiment Analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 9 publications
0
4
0
Order By: Relevance
“…Zhang et al [115] propose a disentangled sentiment representation adversarial network (DiSRAN) to reduce the domain shift of expressive styles for cross-domain sentiment analysis. Recent works [116], [117], [118], [119], [120] tend to focus on disentangling the rich information among multi modalities and leveraging that to perform various downstream tasks. Alaniz et al [116] propose to use the semantic structure of the text to disentangle the visual data, in order to learn an unified representation between the text and image.…”
Section: Multimodal Applicationmentioning
confidence: 99%
See 1 more Smart Citation
“…Zhang et al [115] propose a disentangled sentiment representation adversarial network (DiSRAN) to reduce the domain shift of expressive styles for cross-domain sentiment analysis. Recent works [116], [117], [118], [119], [120] tend to focus on disentangling the rich information among multi modalities and leveraging that to perform various downstream tasks. Alaniz et al [116] propose to use the semantic structure of the text to disentangle the visual data, in order to learn an unified representation between the text and image.…”
Section: Multimodal Applicationmentioning
confidence: 99%
“…Multimodal Application [113], [115] VAE-based Supervised [118] GAN-based [117], [119], [120] Others [26], [28], [123], [124], [125] VAE-based Supervised Recommendation [27], [29], [122] Others [128], [129], [130], [131] VAE-based Supervised Graph [29], [30], [132] Others tation z, which is fine-grained and ii) vector-wise: use two or more independent vectors z 1 , z 2 ... to represent different parts of data features, which is coarse-grained. To guarantee the disentanglement property, approach i) usually requires that z is dimension-wise independent, while approach ii) usually requires that z i is independent with z j where i = j.…”
Section: Papersmentioning
confidence: 99%
“…To increase the precision and viability of the evaluation of troop equipment, they suggested a sentiment analysis model based on attention and contextual factors for extracting essential insights from soldiers' remarks on equipment [3]. Wenwen Zou and others proposed an external hierarchical fusion structure that fuses BERT middle layer information with non-linguistic modal multi-stage fusion for BERT fine-tuning of multimodal linguistic data on CMU-MOSI and CMU-MOSEI datasets [4]. Yusuf Ziya Poyraz explained how BERT was utilized to analyze Sindhi newspaper datasets, which addressed the difficulties in Sindhi language analysis by improving accuracy through transliteration and lexical addition [5].…”
Section: Introductionmentioning
confidence: 99%
“…Further, they can be used to generate novel examples not found in the original dataset [ 21 ]. Such feature learning also supports a variety of other applications, such as super-resolution [ 22 ], multimodal application [ 23 , 24 , 25 , 26 , 27 ], medical imaging [ 28 , 29 ], video prediction [ 30 , 31 , 32 , 33 , 34 ], natural language processing [ 35 , 36 , 37 ], transfer learning and zero-shot learning [ 38 ].…”
Section: Introductionmentioning
confidence: 99%