2017 IEEE International Conference on Computer Vision (ICCV) 2017
DOI: 10.1109/iccv.2017.304
|View full text |Cite
|
Sign up to set email alerts
|

Least Squares Generative Adversarial Networks

Abstract: Unsupervised learning with generative adversarial networks (GANs) has proven hugely successful. Regular GANs hypothesize the discriminator as a classifier with the sigmoid cross entropy loss function. However, we found that this loss function may lead to the vanishing gradients problem during the learning process. To overcome such a problem, we propose in this paper the Least Squares Generative Adversarial Networks (LS-GANs) which adopt the least squares loss function for the discriminator. We show that minimi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

9
2,627
1
4

Year Published

2017
2017
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 4,382 publications
(2,641 citation statements)
references
References 26 publications
9
2,627
1
4
Order By: Relevance
“…Effectively, this is allowing the discriminator to learn its own domain knowledge, and then provide feedback to update the main model. For this study, we utilized the Least Squares GAN (LSGAN) formulation:argtrueminθNDLADVD=12||ND)(yitalictrue-b22+12||ND)(NG)(x-a22argtrueminθNGLADVG=12||ND)(NG)(x-c22where θND and θNG are the trainable weights parameterizing the discriminator network, N D , and generator network, N G , respectively. LADVD and LADVG are the loss functions to be minimized with respect to θND and θNG.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Effectively, this is allowing the discriminator to learn its own domain knowledge, and then provide feedback to update the main model. For this study, we utilized the Least Squares GAN (LSGAN) formulation:argtrueminθNDLADVD=12||ND)(yitalictrue-b22+12||ND)(NG)(x-a22argtrueminθNGLADVG=12||ND)(NG)(x-c22where θND and θNG are the trainable weights parameterizing the discriminator network, N D , and generator network, N G , respectively. LADVD and LADVG are the loss functions to be minimized with respect to θND and θNG.…”
Section: Methodsmentioning
confidence: 99%
“…The discriminator tries to distinguish ytrue from the data created from the generator. As per suggestion by the LSGAN publication, to minimize the Pearson X 2 divergence, we set a = −1, b = 1, and c = 0.…”
Section: Methodsmentioning
confidence: 99%
“…According to the original GAN, the generator GitalicAB and discriminator DB can be trained by solving the following min–max problem: falseminGitalicABfalsemaxDBscriptLGANfalse(GitalicAB,DB,A,Bfalse)=double-struckExBPBfalse[logDB(xB)false]+double-struckExAPAfalse[log(1DBfalse(GitalicAB(xA)false))false],where GitalicAB is trained to reduce a noise in the low‐dose CT image xA to make it similar to the routine‐dose CT image xB, while DB is trained to discriminate between the denoised CT image GABfalse(xAfalse) and the routine‐dose CT image xB. However, we found that the original adversarial loss is unstable during training process; thus, we changed the log‐likelihood function to a least square loss as in the least squares GAN (LSGAN) . Then, the min–max problem can be changed to the two minimization problems as follows: falseminGitalicABdouble-struckExAPAfalse[false(DB(GABfalse(xAfalse))1false)2false],…”
Section: Theorymentioning
confidence: 99%
“…GANs have been reported to be notoriously hard to train in practice and several techniques have been proposed to alleviate some of the complexities involved in getting them to work including modified objective functions and regularization (Salimans et al, 2016;Mao et al, 2016;Gulrajani et al, 2017). We discuss some of these problems in the following subsection.…”
Section: Generative Adversarial Networkmentioning
confidence: 99%
“…Nowozin et al (2016) show that it is possible to train GANs with a variety of f-divergence measures besides JSD. Wasserstein GANs (WGANs) minimize the earth mover's distance or Wasserstein distance, while Least Squared GANs (LSGANs) (Mao et al, 2016) modifies replaces the log loss with an L2 loss. WGAN-GP (Gulrajani et al, 2017) incorporate a gradient penalty term on the discriminator's loss in the WGAN objective which acts as a regularizer.…”
Section: Generative Adversarial Networkmentioning
confidence: 99%