2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01380
|View full text |Cite
|
Sign up to set email alerts
|

Neural Head Reenactment with Latent Pose Descriptors

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
87
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 109 publications
(87 citation statements)
references
References 25 publications
0
87
0
Order By: Relevance
“…, I (K) } by encoder E n . Notably, data augmentation is also introduced in [6] for learning face reenactment. Different from their goal, our derivation of this space is to assist better feature learning and further representation modularization.…”
Section: Identifying Non-identity Feature Spacementioning
confidence: 99%
“…, I (K) } by encoder E n . Notably, data augmentation is also introduced in [6] for learning face reenactment. Different from their goal, our derivation of this space is to assist better feature learning and further representation modularization.…”
Section: Identifying Non-identity Feature Spacementioning
confidence: 99%
“…However, dedicated device setup and heavily manual work are always needed for generating a realistic avatar and reconstructing the detailed appearance, subtle expressions, and gaze movement of a subject. Recent deep-learning based methods [6,12,30,32,58,78,79,82,85,86,88] avoid 3D avatar modeling and directly synthesize a talking head video of a subject from one source image of the subject and a video sequence. Elgharib et al [18] developed a solution for warping the video of a subject's face from side view to front view.…”
Section: Free Viewpoint Video Of Human Charactersmentioning
confidence: 99%
“…However, existing datasets collected from unlisted sources online remain unorganized and noisy, narrowing their applicability for developing data-driven models. For example, previous work on head reenactment [7,17,32,37,41,42] in recent years requires two frames extracted from the same video for training, which is infeasible using existing animation datasets.…”
Section: Related Workmentioning
confidence: 99%