2017 12th IEEE International Conference on Automatic Face &Amp; Gesture Recognition (FG 2017) 2017
DOI: 10.1109/fg.2017.13
|View full text |Cite
|
Sign up to set email alerts
|

Learning Spatial and Temporal Cues for Multi-Label Facial Action Unit Detection

Abstract: Facial action units (AUs) are essential to decode human facial expressions. Researchers have focused on training AU detectors with a variety of features and classifiers. However, several issues remain. These are spatial representation, temporal modeling, and AU correlation. Unlike most studies that tackle these issues separately, we propose a hybrid network architecture to jointly address them. Specifically, spatial representations are extracted by a Convolutional Neural Network (CNN), which, as analyzed in th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
100
0
1

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 134 publications
(101 citation statements)
references
References 41 publications
0
100
0
1
Order By: Relevance
“…In recent years, deep convolutional neural networks have been widely used in a variety of computer vision tasks and have achieved unprecedented progress (He et al 2016;Li et al 2017a;Liu et al 2018). There are also attempts to apply deep CNN to facial AU recognition (Zhao, Chu, and Zhang 2016;Li, Abtahi, and Zhu 2017;Bishay and Patras 2017;Chu, De la Torre, and Cohn 2016). Zhao et al (Zhao, Chu, and Zhang 2016) proposed a unified architecture for facial AU detection which incorporates a deep regional feature learning and multi-label learning modules.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In recent years, deep convolutional neural networks have been widely used in a variety of computer vision tasks and have achieved unprecedented progress (He et al 2016;Li et al 2017a;Liu et al 2018). There are also attempts to apply deep CNN to facial AU recognition (Zhao, Chu, and Zhang 2016;Li, Abtahi, and Zhu 2017;Bishay and Patras 2017;Chu, De la Torre, and Cohn 2016). Zhao et al (Zhao, Chu, and Zhang 2016) proposed a unified architecture for facial AU detection which incorporates a deep regional feature learning and multi-label learning modules.…”
Section: Related Workmentioning
confidence: 99%
“…shape or appearance features) or more effective discriminative learning methods (Valstar and Pantic 2006;Zhao et al 2015;Jiang, Valstar, and Pantic 2011). In recent years, deep convolutional neural networks have been widely used in AU recognition due to their powerful feature representation and end-to-end efficient learning scheme, and have greatly promoted the development of this field (Zhao, Chu, and Zhang 2016;Li, Abtahi, and Zhu 2017;Bishay and Patras 2017;Chu, De la Torre, and Cohn 2016). However, recent efforts based on deep convolutional neural networks are indulged in designing deeper and more complex network structures without exception, learning more robust feature representa-tions in a data-driven manner without explicitly considering and modeling the local characteristics of the facial organs and the linkage relationship between the facial muscles.…”
Section: Introductionmentioning
confidence: 99%
“…Note that LSTM-extended version of ROI [9] is not compared due to its input of a sequence of images instead of a single image. For a fair comparison, we use the results of LSVM, JPML, APL, and CPM reported in [31,3,10]. Table 1.…”
Section: Comparison With State-of-the-art Methodsmentioning
confidence: 99%
“…We collect the F1 scores of the most popular state-of-theart approaches that used the same 3-fold protocol in Table 4 and Table 7 to compare our approaches with other methods. These techniques include a linear support vector machine (LSVM), active patch learning (APL [43]), JPML [13], a confidence-preserving machine (CPM [10]), a block-based region learning CNN (DRML [14]), an enhancing and cropping nets (EAC-net [4]), an ROI adaption net (ROI-Nets [3]), and LSTM fused with a simple CNN (CNN+LSTM [38]), an optimized filter size CNN (OFS-CNN [15]). We also conduct complete control experiments of AU R-CNN in Table 5 and Table 8, Table 3.…”
Section: Evaluation Metricmentioning
confidence: 99%