2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00522
|View full text |Cite
|
Sign up to set email alerts
|

SketchyCOCO: Image Generation From Freehand Scene Sketches

Abstract: We introduce the first method for automatic image generation from scene-level freehand sketches. Our model allows for controllable image generation by specifying the synthesis goal via freehand sketches. The key contribution is an attribute vector bridged Generative Adversarial Network called EdgeGAN, which supports high visual-quality object-level image content generation without using freehand sketches as training data. We have built a largescale composite dataset called SketchyCOCO to support and evaluate t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
88
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 115 publications
(88 citation statements)
references
References 36 publications
0
88
0
Order By: Relevance
“…Furthermore, even artificial sketch generators, when trained to create a sketch to convey the essence of an image with as few strokes as possible learn to first draw lines that have the most power to convey essential content [22]. In fact, there have recently been a number of artificial neural networks trained to generate sketches that are as easily recognizable as those generated by a human [23,24]. We here show that drawings that take advantage of the visual system's mechanisms for understanding scenes will be more easily interpreted [25].…”
Section: Discussionmentioning
confidence: 99%
“…Furthermore, even artificial sketch generators, when trained to create a sketch to convey the essence of an image with as few strokes as possible learn to first draw lines that have the most power to convey essential content [22]. In fact, there have recently been a number of artificial neural networks trained to generate sketches that are as easily recognizable as those generated by a human [23,24]. We here show that drawings that take advantage of the visual system's mechanisms for understanding scenes will be more easily interpreted [25].…”
Section: Discussionmentioning
confidence: 99%
“…It is measured by retrieving relevant text given an image query. For sketch-based image synthesis, classification accuracy is used to measure the realism of the synthesized objects [7,8] and how well the identities of synthesized results match those of real images [26]. Also, similarity between input sketches and edges of synthesized images can be measured to evaluate the correspondence between the input and output [8].…”
Section: How Do We Evaluate the Output Synthesized Images?mentioning
confidence: 99%
“…For sketch-based image synthesis, classification accuracy is used to measure the realism of the synthesized objects [7,8] and how well the identities of synthesized results match those of real images [26]. Also, similarity between input sketches and edges of synthesized images can be measured to evaluate the correspondence between the input and output [8]. In the scenario of pose-guided person image synthesis, "masked" versions of IS and SSIM, Mask-IS and Mask-SSIM are often used to ignore the effects of the background [27][28][29][30][31], since we want to focus on the synthesized human body.…”
Section: How Do We Evaluate the Output Synthesized Images?mentioning
confidence: 99%
See 1 more Smart Citation
“…Besides, Sun et al [31] focused on the layout-to-image generation and proposed an intuitive paradigm to bridge the gap between input labels and generated images. Given a scene sketch, Gao et al [7] implemented a controllable image generation method to meet the specific requirements. Also, there are many generative models [18,9,28,25] that take the text as the input for multi-modal text-to-image generation.…”
Section: 2mentioning
confidence: 99%