2022
DOI: 10.1016/j.knosys.2021.107598
|View full text |Cite
|
Sign up to set email alerts
|

Exploration meets exploitation: Multitask learning for emotion recognition based on discrete and dimensional models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 32 publications
(9 citation statements)
references
References 43 publications
0
9
0
Order By: Relevance
“…Applications such as social networks and news websites have generated abundant multimodal data. As a result, people have conducted extensive multimodal investigations; for example, sentiment analysis [23][24][25][26][27], image and text retrieval [28], reason extraction [29,30], and sarcasm detection [31]. Unlike text-modal-based sarcasm detection, multimodal sarcasm detection aims to identify sarcastic expressions implied in multimodal data.…”
Section: Multimodal Sarcasm Detectionmentioning
confidence: 99%
“…Applications such as social networks and news websites have generated abundant multimodal data. As a result, people have conducted extensive multimodal investigations; for example, sentiment analysis [23][24][25][26][27], image and text retrieval [28], reason extraction [29,30], and sarcasm detection [31]. Unlike text-modal-based sarcasm detection, multimodal sarcasm detection aims to identify sarcastic expressions implied in multimodal data.…”
Section: Multimodal Sarcasm Detectionmentioning
confidence: 99%
“…Among the existing models, the circumplex model is most commonly used in affective computing. This approach easily estimates the similarities and differences between emotions because of the continuity of numerical vectors [12]. In the dimensional approach, emotions are conceptualized within two dimensions, namely arousal, and valence.…”
Section: Introductionmentioning
confidence: 99%
“…Despite the progress, text alone cannot provide sufficient cues for deeper feelings compared to multimodal percep- tion (Hazarika et al 2018). Existing multimodal ERC methods mainly focus on aggregation-based fusion by concatenation (Tu et al 2022b), tensor product (Mai, Hu, and Xing 2019;Liu et al 2018), attention network (Rahman et al 2020;Wang et al 2019) or heterogeneous graph (Yang et al 2021;Hu et al 2022), etc. For instance, Hazarika et al (2018) proposed a conversational memory network to align features from multiple views.…”
Section: Introductionmentioning
confidence: 99%