2019 IEEE/CVF International Conference on Computer Vision (ICCV) 2019
DOI: 10.1109/iccv.2019.00955
|View full text |Cite
|
Sign up to set email alerts
|

Few-Shot Adversarial Learning of Realistic Neural Talking Head Models

Abstract: SourceTarget → Landmarks → Result Source Target → Landmarks → Result Figure 1: The results of talking head image synthesis using face landmark tracks extracted from a different video sequence of the same person (on the left), and using face landmarks of a different person (on the right). The results are conditioned on the landmarks taken from the target frame, while the source frame is an example from the training set. The talking head models on the left were trained using eight frames, while the models on the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
504
0
6

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 611 publications
(511 citation statements)
references
References 33 publications
1
504
0
6
Order By: Relevance
“…This work greatly simplified the requirements for attackers: sim-ply acquire a picture of the victim (usually a profile picture on a social network or an ID photo). Zakharov et al [35] followed up by improving the quality of videos generated using only a few input images. Vougioukas et al [36] raised the bar by introducing a method for animating a facial image from an audio track containing speech.…”
Section: Face Manipulationmentioning
confidence: 99%
“…This work greatly simplified the requirements for attackers: sim-ply acquire a picture of the victim (usually a profile picture on a social network or an ID photo). Zakharov et al [35] followed up by improving the quality of videos generated using only a few input images. Vougioukas et al [36] raised the bar by introducing a method for animating a facial image from an audio track containing speech.…”
Section: Face Manipulationmentioning
confidence: 99%
“…Most of these applications involve image processing. Although there have been some studies involving video processing, such as video generation [115], video colorization [116], [117], video inpainting [118], motion transfer [119], and facial animation synthesis [120]- [123], the research on video using GANs is limited. In addition, although GANs have been applied to the generation and synthesis of 3D models, such as 3D colorization [124], 3D face reconstruction [125], [126], 3D character animation [127], and 3D textured object generation [128], the results are far from perfect.…”
Section: B Future Opportunitiesmentioning
confidence: 99%
“…In fact, the risk from these systems can be understood as a product of their high data efficiency. Deepfakes that required 1000s of hours of video would be much less disruptive, whereas a recent system was able to base deepfakes on as few as 32 video frames [35]. Data efficiency is a crucial parameter in judging misuse potential.…”
Section: New Actors Access ML Capabilitiesmentioning
confidence: 99%