2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU) 2019
DOI: 10.1109/asru46091.2019.9003838
|View full text |Cite
|
Sign up to set email alerts
|

A Cross-Corpus Study on Speech Emotion Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
20
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 26 publications
(22 citation statements)
references
References 29 publications
0
20
0
Order By: Relevance
“…It was shown that there have been many studies on cross-corpus SER performance, mainly transferring the feature space or fine-tuned high-level features. Milner et al [ 63 ] extracted log-mel filterbanks with 23 dimensions and a pretrained source dataset with BiLSTM and transferred the model to learn the target dataset. Table 10 shows that our fine-tuned method outperforms the existing method in RAVDESS, EmoDB, and SAVEE.…”
Section: Resultsmentioning
confidence: 99%
“…It was shown that there have been many studies on cross-corpus SER performance, mainly transferring the feature space or fine-tuned high-level features. Milner et al [ 63 ] extracted log-mel filterbanks with 23 dimensions and a pretrained source dataset with BiLSTM and transferred the model to learn the target dataset. Table 10 shows that our fine-tuned method outperforms the existing method in RAVDESS, EmoDB, and SAVEE.…”
Section: Resultsmentioning
confidence: 99%
“…[3] introduce a domain-adaptive subspace learning method to reduce the feature space differences between source and target speech. In [4] a bidirectional LSTM with attention mechanism is used for classifying emotions across various corpora. To generalize to emotions across corpora authors suggest to use models trained on out-of-domain data and conduct adaptation to the missing corpus or use domain adversarial training (DAT) [5].…”
Section: Related Workmentioning
confidence: 99%
“…One of the possible reasons why attention models outperform others is that the models learn the biases for a specific task, or group of tasks, leading to improved generalisation. Recently, a sequence and attention-based domain adversarial system was presented in [20] which investigated whether the information in acted datasets can be learnt to benefit emotion prediction for natural datasets and achieved state-of-the-art results.…”
Section: Related Workmentioning
confidence: 99%
“…The utterances are split into a training set of 4290 samples (Sessions 1-4) and a test set of 1241 samples(Session 5). This is referred to as IEM4 in this paper and in [20].…”
Section: Datasetmentioning
confidence: 99%
See 1 more Smart Citation