2022
DOI: 10.1016/j.neucom.2021.12.005
|View full text |Cite
|
Sign up to set email alerts
|

DiverGAN: An Efficient and Effective Single-Stage Framework for Diverse Text-to-Image Generation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 19 publications
(22 citation statements)
references
References 59 publications
1
21
0
Order By: Relevance
“…We therefore carry out a serious of experiments on the DiverGAN generator that was trained on three popular text-to-image data sets (i.e., the CUB bird [3], MS COCO [12] and Multi-Modal CelebA-HQ [4] data sets). The experimental results in the current study represent an improvement in performance and explainability in the analyzed algorithm [2]. Meanwhile, our well-trained classifier achieves impressive classification accuracy (bird: 98.09% and face: 99.16%) on the Good & Bad data set and our proposed semantic-discovery algorithm can lead to a more precise control over the latent space of the DiverGAN model, which validate the effectiveness of our presented methods.…”
Section: Introductionsupporting
confidence: 71%
See 4 more Smart Citations
“…We therefore carry out a serious of experiments on the DiverGAN generator that was trained on three popular text-to-image data sets (i.e., the CUB bird [3], MS COCO [12] and Multi-Modal CelebA-HQ [4] data sets). The experimental results in the current study represent an improvement in performance and explainability in the analyzed algorithm [2]. Meanwhile, our well-trained classifier achieves impressive classification accuracy (bird: 98.09% and face: 99.16%) on the Good & Bad data set and our proposed semantic-discovery algorithm can lead to a more precise control over the latent space of the DiverGAN model, which validate the effectiveness of our presented methods.…”
Section: Introductionsupporting
confidence: 71%
“…Zhang et al [36] developed XMC-GAN which studied contrastive learning in the context of text-to-image generation while producing visually plausible images via a simple single-stage framework. Zhang et al [2] presented an efficient and effective single-stage framework called Diver-GAN which is capable of generating diverse, plausible and semantically-consistent images according to a natural-language description. Note that we adopt the DiverGAN generator to perform comprehensive experiments due to its superior performance on image quality and diversity.…”
Section: Cgan In Text-to-image Generationmentioning
confidence: 99%
See 3 more Smart Citations