2019
DOI: 10.48550/arxiv.1905.08233
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Few-Shot Adversarial Learning of Realistic Neural Talking Head Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
69
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 33 publications
(69 citation statements)
references
References 0 publications
0
69
0
Order By: Relevance
“…Multi-domain image-to-image translation with cGAN: Conditioning GANs has been an active research field ever since Mirza et al [29] first showed that auxiliary information such as target label can be fed into standard GANs [12] to generate images conditioned on the auxiliary information. Various forms of representations such as one-hot encoded target domain label [6,8,17,25,51], source domain in the form of semantic text or layout representa-tions [37,49,18,52], source information in the form of images [20,19,26], geometry captured in the form of sparse landmarks [48] have successfully been applied to control the nature of synthetic data. Popular image-to-image translation works such as [20,22] generate visually pleasing images, and may use reconstruction loss such as cycle consistency to preserve the major attributes of the source image [54].…”
Section: Related Workmentioning
confidence: 99%
“…Multi-domain image-to-image translation with cGAN: Conditioning GANs has been an active research field ever since Mirza et al [29] first showed that auxiliary information such as target label can be fed into standard GANs [12] to generate images conditioned on the auxiliary information. Various forms of representations such as one-hot encoded target domain label [6,8,17,25,51], source domain in the form of semantic text or layout representa-tions [37,49,18,52], source information in the form of images [20,19,26], geometry captured in the form of sparse landmarks [48] have successfully been applied to control the nature of synthetic data. Popular image-to-image translation works such as [20,22] generate visually pleasing images, and may use reconstruction loss such as cycle consistency to preserve the major attributes of the source image [54].…”
Section: Related Workmentioning
confidence: 99%
“…Perceptual loss [21] between Ĵfake and J. Perceptual loss is significant to preserve the identity information and high-level semantic features of the face images. We follow the work of [45] and employ the pre-trained VGG-Face [31] network to extract the features:…”
Section: The Training Process Of the Pas Is Guided By The Following L...mentioning
confidence: 99%
“…To solve this issue, few-shot or even one-shot face reenactment methods have also been developed in the recent work [29,31,33]. Wiles et al [29] propose a model, namely X2Face, that is able to use facial landmarks or audio to drive the input source image to a target expression.…”
Section: Related Workmentioning
confidence: 99%
“…Then it produces an intermediate interpolation map given the target expression to be used for transferring the frontalized face. Zakharov et al [31] present a few-shot learning approach that achieves the face reenactment given a few, or even one, source images. Unlike the X2Face model, their method is able to directly transfer the expression without the intermediate boundary latent space [30] or interpolation map [29].…”
Section: Related Workmentioning
confidence: 99%