2023
DOI: 10.3390/math11132908
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Prompt Learning in Emotion Recognition Using Context and Audio Information

Abstract: Prompt learning has improved the performance of language models by reducing the gap in language model training methods of pre-training and downstream tasks. However, extending prompt learning in language models pre-trained with unimodal data to multimodal sources is difficult as it requires additional deep-learning layers that cannot be attached. In the natural-language emotion-recognition task, improved emotional classification can be expected when using audio and text to train a model rather than only natura… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(1 citation statement)
references
References 23 publications
0
1
0
Order By: Relevance
“…Zhao et al [52] use a pretrained language model in conjunction with a prompt and combine the resulting embeddings with data from other modalities. Jeong et al [19] employ something similar but focus only on the combination of text and audio. However, previous work does not evaluate these techniques in a multilingual setting.…”
Section: Prompt-based Learning For Emotion Classificationmentioning
confidence: 99%
“…Zhao et al [52] use a pretrained language model in conjunction with a prompt and combine the resulting embeddings with data from other modalities. Jeong et al [19] employ something similar but focus only on the combination of text and audio. However, previous work does not evaluate these techniques in a multilingual setting.…”
Section: Prompt-based Learning For Emotion Classificationmentioning
confidence: 99%