2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00916
|View full text |Cite
|
Sign up to set email alerts
|

StarGAN: Unified Generative Adversarial Networks for Multi-domain Image-to-Image Translation

Abstract: Figure 1. Multi-domain image-to-image translation results on the CelebA dataset via transferring knowledge learned from the RaFD dataset. The first and sixth columns show input images while the remaining columns are images generated by StarGAN. Note that the images are generated by a single generator network, and facial expression labels such as angry, happy, and fearful are from RaFD, not CelebA. AbstractRecent studies have shown remarkable success in imageto-image translation for two domains. However, existi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

8
3,324
0
11

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 3,591 publications
(3,343 citation statements)
references
References 28 publications
8
3,324
0
11
Order By: Relevance
“…After training the generator in the GAN framework with the discriminator, we can transfer the entire image dataset to the same batch target t using the generator model G such that the batch effect is equalized/mediated and downstream analysis can then be performed on this generated ideal dataset. (Choi et al, 2017). a) Training procedure for the discriminator D. The discriminator D takes an image (either real or generated) and predicts its source (real/generated), which batch it comes from, and biological features such as cell size.…”
Section: Overall Frameworkmentioning
confidence: 99%
See 2 more Smart Citations
“…After training the generator in the GAN framework with the discriminator, we can transfer the entire image dataset to the same batch target t using the generator model G such that the batch effect is equalized/mediated and downstream analysis can then be performed on this generated ideal dataset. (Choi et al, 2017). a) Training procedure for the discriminator D. The discriminator D takes an image (either real or generated) and predicts its source (real/generated), which batch it comes from, and biological features such as cell size.…”
Section: Overall Frameworkmentioning
confidence: 99%
“…Since our framework heavily relies on the original StarGAN (Choi et al, 2017), we will first do a brief review of the StarGAN formulation, and introduce our extension for representation disentanglement in the following section. Under the GAN framework, our discriminator D will learn to predict whether an image is a real image from the original data distribution or a generated image from G. This is the image source prediction D src (X).…”
Section: Stargan For Batch Transformationmentioning
confidence: 99%
See 1 more Smart Citation
“…Our approach builds on the CycleGAN framework [ZPIE17] that learns a mapping function between two unpaired image domains using two GAN models [GPAM*14]. The works of [LZZ16, SLH*17, CCK*17] developed variants of GAN models for face attributes synthesis, such as gender/age modification and expression transformation. However, those imaged‐based GAN models are usually limited to generating low‐resolution images and likely to incur artefacts on face or background change.…”
Section: Related Workmentioning
confidence: 99%
“…CycleGAN uses a cycle-consistency learning technique to translate images from one domain to the other one and exhibits comparable performance to supervised methods. Such learning strategy and its modified versions have been validated in style transfer of natural images [28][29][30] and medical image analysis [31][32][33] . As for the field of optical microscopy, a few forward-looking studies have applied cycleGANs to remove coherent noise in optical diffraction tomography 34 , and realize image segmentation for bright-field microscopy and X-ray computed tomography 35 .…”
mentioning
confidence: 99%