2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00618
|View full text |Cite
|
Sign up to set email alerts
|

StyleRig: Rigging StyleGAN for 3D Control Over Portrait Images

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
239
2

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 376 publications
(241 citation statements)
references
References 24 publications
0
239
2
Order By: Relevance
“…Deng et al [2020] imitate the 3D rendering process and introduce contrastive learning to learn a disentangled latent space. Many other works (e.g., Härkönen et al 2020;Shen et al 2020;Tewari et al 2020]) have tried to analyze and disentangle the latent code of some pretrained GAN space ] also with labeled data of specific attributes. Although these works successfully disentangle the latent space, they could only control a limited number of predefined attributes such as gender, expression, and age, due to the use of labeled data in the training stage.…”
Section: Neural Image Disentanglementmentioning
confidence: 99%
“…Deng et al [2020] imitate the 3D rendering process and introduce contrastive learning to learn a disentangled latent space. Many other works (e.g., Härkönen et al 2020;Shen et al 2020;Tewari et al 2020]) have tried to analyze and disentangle the latent code of some pretrained GAN space ] also with labeled data of specific attributes. Although these works successfully disentangle the latent space, they could only control a limited number of predefined attributes such as gender, expression, and age, due to the use of labeled data in the training stage.…”
Section: Neural Image Disentanglementmentioning
confidence: 99%
“…The work of Shen et al [5] shows that GANs trained on high quality images learn various semantics in some linear subspaces of the latent space. Tewari et al [18] introduced an approach that provides a face rig-like control on generated images by training a rigging network between 3D morphable face model's semantic parameters and StyleGAN's input. Other approaches have attempted to imitate or directly carry out Principal Component Analysis (PCA) in the latent space of generative networks [6,19].…”
Section: Related Workmentioning
confidence: 99%
“…We build on the StyleGAN architecture [Karras et al 2019[Karras et al , 2020b] that is inherently a 2D image generator. Recent methods [Abdal et al 2021;Tewari et al 2020a] can render head poses parameterized by two angles, but have no way to generate an image for a specific and complete 3D camera, ignoring at least five degrees of freedom (DoF). This results in a small subspace of 3D camera poses; the nature and limits of this subspace have never been precisely defined, much less extended.…”
Section: Introductionmentioning
confidence: 99%