2023
DOI: 10.48550/arxiv.2303.05134
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

hierarchical network with decoupled knowledge distillation for speech emotion recognition

Abstract: The goal of Speech Emotion Recognition (SER) is to enable computers to recognize the emotion category of a given utterance in the same way that humans do. The accuracy of SER is strongly dependent on the validity of the utterance-level representation obtained by the model. Nevertheless, the "dark knowledge" carried by non-target classes is always ignored by previous studies. In this paper, we propose a hierarchical network, called DKDFMH, which employs decoupled knowledge distillation in a deep convolutional n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 24 publications
0
1
0
Order By: Relevance
“…Tang et al [32] distill knowledge from the BERT model [33] into a single-layer bidirectional long short-term memory for natural language processing sentiment classification. DKDFMH [34] uses a fused multihead attention mechanism to employ decoupled knowledge distillation in CNNs, helping the method focus on the distinctions between sentiment features.…”
Section: Knowledge Distillation For Sentimentsmentioning
confidence: 99%
“…Tang et al [32] distill knowledge from the BERT model [33] into a single-layer bidirectional long short-term memory for natural language processing sentiment classification. DKDFMH [34] uses a fused multihead attention mechanism to employ decoupled knowledge distillation in CNNs, helping the method focus on the distinctions between sentiment features.…”
Section: Knowledge Distillation For Sentimentsmentioning
confidence: 99%