2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.00669
|View full text |Cite
|
Sign up to set email alerts
|

Learning a Facial Expression Embedding Disentangled from Identity

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 61 publications
(25 citation statements)
references
References 28 publications
0
25
0
Order By: Relevance
“…Accuracy(%) Average Accuracy(%) DLP-CNN [43] 84.22 74.20 FSN [51] 81.10 72.46 paCNN [45] 83.27 -gaCNN [47] 85.07 -IPA2LT [46] 86.77 -ALT [52] 84.50 76.50 separate loss [34] 86.38 77.25 WS-LGAN [53] 85.07 -RAN [48] 86.90 -DLN [18] 86.4 -SCN [49] 87.03 -Ad-Corre 86.96 79.01 Disgust and Anger (about 13%), while most confusion, about 18%, occurred between Sad and Neutral. In addition, we report the precision, recall, and f1-score of our proposed Ad-Corre Loss on AffectNet [44], RAF-DB [43] and FER-2013 [42] datsets in Tables 5, 6, 7 respectively.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Accuracy(%) Average Accuracy(%) DLP-CNN [43] 84.22 74.20 FSN [51] 81.10 72.46 paCNN [45] 83.27 -gaCNN [47] 85.07 -IPA2LT [46] 86.77 -ALT [52] 84.50 76.50 separate loss [34] 86.38 77.25 WS-LGAN [53] 85.07 -RAN [48] 86.90 -DLN [18] 86.4 -SCN [49] 87.03 -Ad-Corre 86.96 79.01 Disgust and Anger (about 13%), while most confusion, about 18%, occurred between Sad and Neutral. In addition, we report the precision, recall, and f1-score of our proposed Ad-Corre Loss on AffectNet [44], RAF-DB [43] and FER-2013 [42] datsets in Tables 5, 6, 7 respectively.…”
Section: Methodsmentioning
confidence: 99%
“…Dharanya et al [16] proposed Auxiliary Classifier Generative Adversarial Network (AC-GAN) based model which regenerates the basic facial emotions from an input face image and then classifies them. Zhang et al [18] proposed a weakly supervised local-global attention network which is designed to extract and combine the local and the global features from input facial images. Also, their proposed architecture is designed to use the attention mechanism to deal with part location and feature fusion problems.…”
Section: Related Workmentioning
confidence: 99%
“…Due to the limited subjects and unbalanced annotations of existed affective datasets, it is a challenging issue to prevent the emotion recognition model from overfitting on the disturbing factors, like background or random noise. To tackle this problem, we adopt a prior facial expression embedding model [13], which can capture the detailed expression similarities across different people, into our framework. The expression embedding brings at least two advantages.…”
Section: Overviewmentioning
confidence: 99%
“…First, by training on even larger facial image datasets with the identity invariant constraint, the embedding itself is independent to the identity attributes and therefore can improve the network's generalizability to unseen subjects. Second, the expression embedding model [13] is targeted for discriminating the minor expression similarities within triplet training data. It provides a nice initialization for our latter emotion recognition tasks.…”
Section: Overviewmentioning
confidence: 99%
See 1 more Smart Citation