2020
DOI: 10.1007/978-3-030-58542-6_24
|View full text |Cite
|
Sign up to set email alerts
|

SESAME: Semantic Editing of Scenes by Adding, Manipulating or Erasing Objects

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
52
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
5

Relationship

3
7

Authors

Journals

citations
Cited by 68 publications
(52 citation statements)
references
References 36 publications
0
52
0
Order By: Relevance
“…As image inpainting requires a high-level semantic context, and to explicitly include it in the generation pipeline, there exist hand-crafted architectural designs such as Dilated Convolutions [13,38] to increase the receptive field, Partial Convolutions [16] and Gated Convolutions [41] to guide the convolution kernel according to the inpainted mask, Contextual Attention [39] to leverage on global information, Edges maps [7,22,36,37] or Semantic Segmentation maps [11,25] to further guide the generation, and Fourier Convolutions [32] to include both global and local information efficiently. Although recent works produce photo-realistic results, GANs are well known for textural synthesis, so these methods shine on background completion or removing objects, which require repetitive structural synthesis, and struggle with semantic synthesis (See Figure 5).…”
Section: Related Workmentioning
confidence: 99%
“…As image inpainting requires a high-level semantic context, and to explicitly include it in the generation pipeline, there exist hand-crafted architectural designs such as Dilated Convolutions [13,38] to increase the receptive field, Partial Convolutions [16] and Gated Convolutions [41] to guide the convolution kernel according to the inpainted mask, Contextual Attention [39] to leverage on global information, Edges maps [7,22,36,37] or Semantic Segmentation maps [11,25] to further guide the generation, and Fourier Convolutions [32] to include both global and local information efficiently. Although recent works produce photo-realistic results, GANs are well known for textural synthesis, so these methods shine on background completion or removing objects, which require repetitive structural synthesis, and struggle with semantic synthesis (See Figure 5).…”
Section: Related Workmentioning
confidence: 99%
“…The arise of generative adversarial network (GANs) brings revolutionary advance to image editing [4], [20], [52], [55], [67], [68], [94]. As one of the most intuitive representation in image editing, semantic information has been extensively investigated in conditional image synthesis.…”
Section: Semantic Image Editingmentioning
confidence: 99%
“…Conditional normalization is the workhorse of many methods solving diverse problems in multi-domain learning [46,47], image generation [23,9,43], image editing [42], style transfer [19], and super-resolution [56]. The operating principle is as simple as applying condition-dependent affine transformations on the normalized batch [21], local response [26], instance [52], layer [4], or feature group [58], allowing features to occupy different regions in the space, depending on the triggered condition.…”
Section: Conditional Normalization (Cn)mentioning
confidence: 99%