2016
DOI: 10.48550/arxiv.1610.02454
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning What and Where to Draw

Abstract: Generative Adversarial Networks (GANs) have recently demonstrated the capability to synthesize compelling real-world images, such as room interiors, album covers, manga, faces, birds, and flowers. While existing models can synthesize images based on global constraints such as a class label or caption, they do not provide control over pose or object location. We propose a new model, the Generative Adversarial What-Where Network (GAWWN), that synthesizes images given instructions describing what content to draw … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 5 publications
0
9
0
Order By: Relevance
“…Over the next few years, progress continued using a combination of methods. These include improving the generative model architecture with modifications like multi-scale generators (Zhang et al, 2017;, integrating attention and auxiliary losses (Xu et al, 2018), and leveraging additional sources of conditioning information beyond just text (Reed et al, 2016a;Li et al, 2019;Koh et al, 2021).…”
Section: Introductionmentioning
confidence: 99%
“…Over the next few years, progress continued using a combination of methods. These include improving the generative model architecture with modifications like multi-scale generators (Zhang et al, 2017;, integrating attention and auxiliary losses (Xu et al, 2018), and leveraging additional sources of conditioning information beyond just text (Reed et al, 2016a;Li et al, 2019;Koh et al, 2021).…”
Section: Introductionmentioning
confidence: 99%
“…In recent times, GAN-based approaches have been used to generate impressively realistic house-numbers [4], faces, bedrooms [17] and a variety of other image categories [18,21]. Usually, these image categories tend to have extremely complex underlying distributions.…”
Section: Introductionmentioning
confidence: 99%
“…However, their training is notoriously difficult and requires much effort to create any satisfactory results. GANs take as an input a vector, which might be random, fixed, or generated from the text [29,30,31].…”
Section: Generative Adversarial Networkmentioning
confidence: 99%