2018
DOI: 10.48550/arxiv.1807.00751
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Understanding the Effectiveness of Lipschitz-Continuity in Generative Adversarial Nets

Abstract: In this paper, we investigate the underlying factor that leads to failure and success in the training of GANs. We study the property of the optimal discriminative function and show that in many GANs, the gradient from the optimal discriminative function is not reliable, which turns out to be the fundamental cause of failure in the training of GANs. We further demonstrate that a well-defined distance metric does not necessarily guarantee the convergence of GANs. Finally, we prove in this paper that Lipschitz-co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 23 publications
0
8
0
Order By: Relevance
“…Thus, a typical strategy is to enforce K-Lipschitz constraint. Yet, in context of DL it is still unclear if and how it is possible to enforce a model to be exact K-Lipschitz, even though there are several recently proposed techniques for this (Gulrajani et al, 2017;Petzka et al, 2017;Miyato et al, 2018;Zhou et al, 2018).…”
Section: Pso With Unit Magnitudesmentioning
confidence: 99%
“…Thus, a typical strategy is to enforce K-Lipschitz constraint. Yet, in context of DL it is still unclear if and how it is possible to enforce a model to be exact K-Lipschitz, even though there are several recently proposed techniques for this (Gulrajani et al, 2017;Petzka et al, 2017;Miyato et al, 2018;Zhou et al, 2018).…”
Section: Pso With Unit Magnitudesmentioning
confidence: 99%
“…We test the effects of training stabilization brought by the SsGAN. We consider two types of hyper-parameter settings: First, controlling the Lipschitz constant of the discriminator, which is a central quantity analyzed in the GAN literature [12,32]. We consider two state-of-theart techniques: Gradient Penalty [11], and Spectral Normalization [12].…”
Section: Robustness Testmentioning
confidence: 99%
“…The proposition shows that the optimal f * provides informative gradient [57] from q towards p r . We then generalize the conclusion to p θ by considering correlation between q and p θ .…”
Section: Theoretical Analysismentioning
confidence: 92%
“…Research on the Lipschitz continuity of GAN discriminators have resulted in the theory of "informative gradients" [56,57]. Under certain mild conditions, a Lipschitz discriminator can provide informative gradient to the generator in a GAN framework: when p θ and p r are disjoint, the gradient ∇f * (x) of optimal discriminator f * w.r.t each sample x ∼ p θ points to a sample x * ∼ p r , which guarantees that the generation distribution p θ is moving towards p r .…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation