2020
DOI: 10.48550/arxiv.2003.07449
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Object-Centric Image Generation from Layouts

Abstract: Despite recent impressive results on single-object and singledomain image generation, the generation of complex scenes with multiple objects remains challenging. In this paper, we start with the idea that a model must be able to understand individual objects and relationships between objects in order to generate complex scenes well. Our layoutto-image-generation method, which we call Object-Centric Generative Adversarial Network (or OC-GAN), relies on a novel Scene-Graph Similarity Module (SGSM). The SGSM lear… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
36
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(36 citation statements)
references
References 43 publications
0
36
0
Order By: Relevance
“…The final images are then produced via a series of convolutional "refinement" layers [13,2] or a pre-trained SPADE [21] model. In contrast, [26,27,28] generate scenes in a GAN setting. The generator is conditioned on the scene layout and mask predictors via a layout-aware norm, inspired by StyleGAN [15].…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…The final images are then produced via a series of convolutional "refinement" layers [13,2] or a pre-trained SPADE [21] model. In contrast, [26,27,28] generate scenes in a GAN setting. The generator is conditioned on the scene layout and mask predictors via a layout-aware norm, inspired by StyleGAN [15].…”
Section: Related Workmentioning
confidence: 99%
“…The generator is conditioned on the scene layout and mask predictors via a layout-aware norm, inspired by StyleGAN [15]. [28] further introduce a differentiable graph-scene matching loss to enhance the separation of objects. Orthogonal to these approaches, we avoid intermediate steps and after compressing high-frequency image details (e.g.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Object-level image-to-image translation. The objectlevel I2I translation is derived from the development of object-driven image generation research, e.g., synthesizing images from object scenes [1,12] and synthesizing images from scene layouts [32,34,39]. The object-level I2I translation methods utilize object perception (bounding boxes or masks) as guidance, which is effective for generating sharp object boundaries.…”
Section: Related Workmentioning
confidence: 99%