2017
DOI: 10.48550/arxiv.1704.00028
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Improved Training of Wasserstein GANs

Abstract: Generative Adversarial Networks (GANs) are powerful generative models, but suffer from training instability. The recently proposed Wasserstein GAN (WGAN) makes progress toward stable training of GANs, but sometimes can still generate only poor samples or fail to converge. We find that these problems are often due to the use of weight clipping in WGAN to enforce a Lipschitz constraint on the critic, which can lead to undesired behavior. We propose an alternative to clipping weights: penalize the norm of gradien… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

7
928
2

Year Published

2021
2021
2022
2022

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 523 publications
(1,010 citation statements)
references
References 22 publications
7
928
2
Order By: Relevance
“…We train our denoising diffusion GAN with 4 denoising steps and compare it to other models in Figure 6. We observe that the vanilla GAN suffers severely from mode collapse, and while techniques like WGAN-GP (Gulrajani et al, 2017) improve mode coverage, the sample quality is still limited.…”
Section: Additional Studiesmentioning
confidence: 95%
“…We train our denoising diffusion GAN with 4 denoising steps and compare it to other models in Figure 6. We observe that the vanilla GAN suffers severely from mode collapse, and while techniques like WGAN-GP (Gulrajani et al, 2017) improve mode coverage, the sample quality is still limited.…”
Section: Additional Studiesmentioning
confidence: 95%
“…For the perceptual loss, we calculate it on relu5-1 VGG (Simonyan and Zisserman 2014) features. For the adversarial loss, we employ WGAN-GP (Gulrajani et al 2017). The overall loss function of our model is ultimately designed as:…”
Section: Loss Functionsmentioning
confidence: 99%
“…The WGAN framework requires the discriminator to be Lipschitz continuous with respect to the input. Therefore, we introduce a gradient penalty term to constrain the gradient of the discriminator [33]:…”
Section: Generative Adversarial Neural Networkmentioning
confidence: 99%