2017
DOI: 10.48550/arxiv.1710.10196
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Progressive Growing of GANs for Improved Quality, Stability, and Variation

Abstract: We describe a new training methodology for generative adversarial networks. The key idea is to grow both the generator and discriminator progressively: starting from a low resolution, we add new layers that model increasingly fine details as training progresses. This both speeds the training up and greatly stabilizes it, allowing us to produce images of unprecedented quality, e.g., CELEBA images at 1024 2 . We also propose a simple way to increase the variation in generated images, and achieve a record incepti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

5
1,841
0
6

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
4

Relationship

0
10

Authors

Journals

citations
Cited by 1,167 publications
(1,852 citation statements)
references
References 11 publications
5
1,841
0
6
Order By: Relevance
“…StyleGAN [24] combines the architecture of Progressive Growing GAN [25] with the style transfer principles [26] in Figure 2. StyleGAN's architecture addresses some limitations of the GAN models, such as stability during training and lack of control over the images generated.…”
Section: Theoretical Background 21 Styleganmentioning
confidence: 99%
“…StyleGAN [24] combines the architecture of Progressive Growing GAN [25] with the style transfer principles [26] in Figure 2. StyleGAN's architecture addresses some limitations of the GAN models, such as stability during training and lack of control over the images generated.…”
Section: Theoretical Background 21 Styleganmentioning
confidence: 99%
“…CelebAMask-HQ [17] comes closest to meeting these requirements among the many face segmentation datasets. It provides manually annotated masks over 30K high-resolution face images in CelebA-HQ [13], covering 18 facial attributes. Nevertheless, this dataset was designed for a GAN-based face image editing task, thus ignoring the various occlusions on the face image (except for hair and glasses), e.g., it wrongly annotates occlusions such as hands as attributes that they obscure.…”
Section: Introductionmentioning
confidence: 99%
“…Since a goal of GAN is discovering the Nash equilibrium of non-convex game in the high-dimensional parameter space, GAN is substantially more complex and difficult to train compared to networks trained by supervised learning [49]. To address this issue, some papers [4,15,48] investigated novel network architectures for discriminator and generator. Although these methods can produce high-resolution images on challenging datasets such as ImageNet [20], they still have the fundamental problem related to the training instability.…”
Section: Introductionmentioning
confidence: 99%