2018
DOI: 10.48550/arxiv.1811.04544
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Visual Saliency Maps Can Apply to Facial Expression Recognition

Zhenyue Qin,
Jie Wu

Abstract: Human eyes concentrate different facial regions during distinct cognitive activities. We study utilising facial visual saliency maps to classify different facial expressions into different emotions. Our results show that our novel method of merely using facial saliency maps can achieve a descent accuracy of 65%, much higher than the chance level of 1/7. Furthermore, our approach is of semi-supervision, i.e., our facial saliency maps are generated from a general saliency prediction algorithm that is not explici… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 27 publications
0
5
0
Order By: Relevance
“…Specifically, first cut the upper left corner, lower left corner, upper right corner, lower right corner and center of the image and then perform horizontal flipping operations on them, respectively, expanding the number of datasets by a factor of 10. That is, for an image, the 10 images obtained after processing are sent to the model, and then the probability values obtained from each image are averaged to obtain 7 average probability values, and the expression class corresponding to the maximum probability value is taken as the final output expression [48].…”
Section: Data Pre-processingmentioning
confidence: 99%
“…Specifically, first cut the upper left corner, lower left corner, upper right corner, lower right corner and center of the image and then perform horizontal flipping operations on them, respectively, expanding the number of datasets by a factor of 10. That is, for an image, the 10 images obtained after processing are sent to the model, and then the probability values obtained from each image are averaged to obtain 7 average probability values, and the expression class corresponding to the maximum probability value is taken as the final output expression [48].…”
Section: Data Pre-processingmentioning
confidence: 99%
“…Mollahosseini et al [21] applied a convolutional neural network to FER. Qin and Wu [22] utilized Resnet [23] to achieve good recognition performance on multiple datasets. Ding et al [24] proposed to train an expression recognition network called the FaceNet2E xpNet based on images.…”
Section: Spatial-based Approachesmentioning
confidence: 99%
“…The experimental results are shown in Figure 7. It turns out Methods Years Accuracy(%) Conv+Inception [30] 2016 66.40 SMFER [31] 2018 65.12 GoogleNet [32] 2018 65.20 VGG+SVM [33] 2019 66.31 SAP [34] 2019 71.08 E-FCNN [28] 2021 66.17 DeepEmotion [35] 2021 uses transfer learning technique to overcome the shortage of training samples. For FER2013, the second-highest recognition accuracy of the test set is SAP [34], which is a sample awareness-based expression recognition method, in which a Bayesian classifier is used to select the most appropriate classifier from a set of candidate classifiers for the current test sample, and then the classifier is used to perform expression recognition on the current sample.…”
Section: E Visualization Analysismentioning
confidence: 99%