2019
DOI: 10.1016/j.neucom.2019.05.003
|View full text |Cite
|
Sign up to set email alerts
|

Conditional adversarial synthesis of 3D facial action units

Abstract: Employing deep learning-based approaches for finegrained facial expression analysis, such as those involving the estimation of Action Unit (AU) intensities, is difficult due to the lack of a large-scale dataset of real faces with sufficiently diverse AU labels for training. In this paper, we consider how AU-level facial image synthesis can be used to substantially augment such a dataset. We propose an AU synthesis framework that combines the well-known 3D Morphable Model (3DMM), which intrinsically disentangle… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5

Relationship

3
2

Authors

Journals

citations
Cited by 24 publications
(16 citation statements)
references
References 37 publications
0
16
0
Order By: Relevance
“…(6) is to alleviate the data imbalance problem. For most facial AU detection benchmarks, the occurrence rates of AUs are imbalanced [13,12]. Since AUs are not mutually independent, imbalanced training data has a bad influence on this multi-label learning task.…”
Section: Facial Au Detectionmentioning
confidence: 99%
“…(6) is to alleviate the data imbalance problem. For most facial AU detection benchmarks, the occurrence rates of AUs are imbalanced [13,12]. Since AUs are not mutually independent, imbalanced training data has a bad influence on this multi-label learning task.…”
Section: Facial Au Detectionmentioning
confidence: 99%
“…Recently, Li et al [10] proposed the AU semantic relationship embedded representation learning (SRERL) framework to combine facial AU detection and Gated Graph Neural Network (GGNN) [12] and achieved good results. But the commonly used Graph Convolutional Network (GCN) for classification task with relation modeling is adopted for AU relation modeling in our proposed method, while the Gated Graph Neural Network (GGNN) adopted in [13] is inspired by GRU and mainly used for the task of Visual Question Answering and Semantic Segmentation. In addition, our method has only about 2.3 million parameters, but SRERL [13] has more than 138 million parameters.…”
Section: Related Workmentioning
confidence: 99%
“…The trade-off weight w i is introduced to alleviate the data imbalance problem. For most facial AU detection benchmarks, the occurrence rates of AUs are imbalanced [13,14]. Since AUs are not mutually independent, imbalanced training data has a bad influence on this multi-label learning task.…”
Section: Facial Au Detectionmentioning
confidence: 99%
“…However, the resolution of its generated facial image is only 32 × 32, and the generated facial images with AU labels are not well quantitatively evaluated. By integrating with the 3D Morphable Model (3DMM), Liu et al [13] proposed an 3D AU synthesis framework to transfer AUs in the range of intensities, however, some texture details are lost for certain AUs because of the limitations of 3D face model. In addition, while generating a single AU, all these above mentioned methods can damage the other AU without keeping the identity information untouched.…”
Section: Facial Expression Synthesismentioning
confidence: 99%
“…It is difficult for the method to synthesize a single AU respectively without keeping the other AU untouched. Liu et al [13] uses 3D Morphable Model (3DMM) to achieve AU synthesis, where the transformation from a image to 3D model damages the texture details of the original images.…”
Section: Introductionmentioning
confidence: 99%