2018
DOI: 10.1002/cav.1819
|View full text |Cite
|
Sign up to set email alerts
|

Emotional facial expression transfer from a single image via generative adversarial nets

Abstract: Facial expression transfer from a single image is a challenging task and has drawn sustained attention in the fields of computer vision and computer graphics. Recently, generative adversarial nets (GANs) have provided a new approach to facial expression transfer from a single image toward target facial expressions.However, it is still difficult to obtain a sequence of smoothly changed facial expressions. We present a novel GAN-based method for generating emotional facial expression animations given a single im… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(11 citation statements)
references
References 23 publications
0
11
0
Order By: Relevance
“…Other alternative models that benefit from a different loss metric are GAN based on Category Information (CIGAN) [57], hinge loss [47], least-square GAN [58], and f-divergence FIGURE 2: Block diagram of LAPGAN model [41]; G k : k-th Generator, D k : k-th Discriminator, X r1 : real sample, X rk : k-th real residual, X gk : generated sample, O 1 : Output of binary classification to real/fake. (a) BiGAN/ALI [61] (b) VAEGAN [62] FIGURE 4: General Structure of GAN models varying by generator; Z: input noise, G:Generator, D:Discriminator, x r : real sample, x g : generated sample, O 1 : Output of binary classification to real/fake, G e : incorporated Generator-encoder, G d : incorporated Generator-decoder, Z r : latent vector that encodes the input for G e , Z g : latent vector that encodes the input for G d .…”
Section: Variants By Discriminatorsmentioning
confidence: 99%
See 1 more Smart Citation
“…Other alternative models that benefit from a different loss metric are GAN based on Category Information (CIGAN) [57], hinge loss [47], least-square GAN [58], and f-divergence FIGURE 2: Block diagram of LAPGAN model [41]; G k : k-th Generator, D k : k-th Discriminator, X r1 : real sample, X rk : k-th real residual, X gk : generated sample, O 1 : Output of binary classification to real/fake. (a) BiGAN/ALI [61] (b) VAEGAN [62] FIGURE 4: General Structure of GAN models varying by generator; Z: input noise, G:Generator, D:Discriminator, x r : real sample, x g : generated sample, O 1 : Output of binary classification to real/fake, G e : incorporated Generator-encoder, G d : incorporated Generator-decoder, Z r : latent vector that encodes the input for G e , Z g : latent vector that encodes the input for G d .…”
Section: Variants By Discriminatorsmentioning
confidence: 99%
“…In 2018, the G2GAN [98] was extended by Qiao et al [62]. The authors derived a model based on VAEGANs to…”
Section: ) Image Synthesismentioning
confidence: 99%
“…More recently, another approach [34] is using fiducial points to geometrically control the face animation while Tulyakov et al [36] is learning to directly generate sequences of images, based on a "content and motion" approach. Quia et al [27] use facial landmarks to improve the animation smoothness of a changing emotion. Kim et al [15] enable to generate video face animation using another portrait video as an example.…”
Section: Related Workmentioning
confidence: 99%
“…For generating artificial representation of emotions, first works from computer graphics community focus on the animation of the faces with model-based approaches [15], [33], [41]. More recently deep learning approaches and especially Generative Adversarial Networks [5], [7], [11], [27], [36] have been proposed, often borrowing ideas from the computer graphics community [26], [34] but also aiming to learn these facial expressions from diverse datasets and representations. Nevertheless, these approaches are generally trained on small corpus with pronounced emotions [19], [20], [43], meaning that the space of possible generated expressions is limited.…”
Section: Introductionmentioning
confidence: 99%
“…Pumarola et al [23] and Shao et al [28] exploited Facial Action Units (AU), and the expression synthesis process is guided by the learned AU features. Similarly, in [30] and [25], facial landmarks are used to produce synthesized expression images.…”
Section: Ter-ganmentioning
confidence: 99%