Proceedings of the 2020 International Conference on Multimodal Interaction 2020
DOI: 10.1145/3382507.3418813
|View full text |Cite
|
Sign up to set email alerts
|

Speaker-Invariant Adversarial Domain Adaptation for Emotion Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 29 publications
(14 citation statements)
references
References 30 publications
0
14
0
Order By: Relevance
“…The best result for the subject-invariant model was obtained with λ = 0.2, and the best result for the subject-invariant and domain adapted model was obtained with λ s = 0.15 and λ d = 0.1 (see Equation 1 and 3). We compared the performance of our method with a EEG CNN encoder [7], DANN [9] and SIDANN [8]. The results are summarized in Table 1.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…The best result for the subject-invariant model was obtained with λ = 0.2, and the best result for the subject-invariant and domain adapted model was obtained with λ s = 0.15 and λ d = 0.1 (see Equation 1 and 3). We compared the performance of our method with a EEG CNN encoder [7], DANN [9] and SIDANN [8]. The results are summarized in Table 1.…”
Section: Resultsmentioning
confidence: 99%
“…The between- database results show that subject-invariant learning, in addition to domain adaptation methods, can partially address the lack of generalization in EEG-base emotion recognition. Our proposed method significantly outperforms SIDANN [8] (significance is tested via a t-test p − value < .05).…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Second, the privacy filter (PF) obfuscates the user's personal information such as face, eye, age, gender, and emotion, using data manipulations such as 'deepfake' approaches [25,33,41] (Figure 1). We select these filters following previous works [1,45] that highlight the risks of video analytics (e.g., emotion recognition [44]) [45] and users concerns in ubiquitous environments regarding the types of collected data [1]. Moreover, these filters are easier for the participants to understand our prototype.…”
Section: System Overviewmentioning
confidence: 99%
“…Although considerable research effort has been put into developing models capable of recognizing and predicting human emotions, much of the work focuses on lab-produced datasets [3] and robust speech emotion recognition on in-thewild speech with diverse confounding acoustic elements remains a challenge [32]. Previous work has used a variety of machine learning techniques such as multi-task learning [32] and domain generalization [31] that attempt to learn a more robust and generalizable representation of speech features, yet few studies have successfully accounted for variations in speech emotion corpora due to the complexity of emotion expressions and a variety of moderating variables such as gender, culture, etc.…”
Section: Introductionmentioning
confidence: 99%