2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2023
DOI: 10.1109/cvpr52729.2023.01761
|View full text |Cite
|
Sign up to set email alerts
|

Imagen Editor and EditBench: Advancing and Evaluating Text-Guided Image Inpainting

Su Wang,
Chitwan Saharia,
Ceslee Montgomery
et al.
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
9
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 57 publications
(9 citation statements)
references
References 28 publications
0
9
0
Order By: Relevance
“…The first set comprises the structure loss [Eq. (11)] and the texture loss [Eq. (12)], aiming at guiding the inference process for the structure and texture of the inpainted regions.…”
Section: Loss Functionmentioning
confidence: 99%
See 1 more Smart Citation
“…The first set comprises the structure loss [Eq. (11)] and the texture loss [Eq. (12)], aiming at guiding the inference process for the structure and texture of the inpainted regions.…”
Section: Loss Functionmentioning
confidence: 99%
“…Furthermore, based on this technology, people can leverage multimodal information to guide the image restoration process. For example, text descriptions or shape constraints can be used to guide the repair of missing areas within the image, 10,11 resulting in more comprehensive and accurate image content restoration.…”
Section: Introductionmentioning
confidence: 99%
“…However, these models are limited in filling in the content using out-of-mask context only. A more flexible usage is adding text control [2,3,60,[66][67][68] that allow for text-buided image inpainting. Latent Blended Diffusion [3] proposed blending the generated and original image latents, Imagenator [60] and Diffusion-based Inpainting [45] fine-tune pre-trained text-to-image generation models with masked images as additional input, and SmartBrush [66] fine-tunes an additional mask prediction branch on object-centric datasets.…”
Section: Related Workmentioning
confidence: 99%
“…[149] Ref. Image, Text ✓ ✓ ✓ ✓ SmartBrush [150] Text, Mask ✓ ✓ ✓ IIR-Net [151] Text ✓ ✓ ✓ PowerPaint [152] Text, Mask ✓ ✓ ✓ Imagen Editor [153] Text, Mask ✓ ✓ ✓ ✓ ✓ SmartMask [154] Text ✓ Uni-paint [155] Text, Mask, Ref. [173] Text, Ref.…”
mentioning
confidence: 99%
“…Anydoor [147], FADING [148], PAIR Diffusion [149], SmartBrush [150], IIR-Net [151], PowerPaint [152], Imagen Editor [153], SmartMask [154], Uni-paint [155] Instructional Editing via Full Supervision InstructPix2Pix [156], MoEController [157], FoI [158], LOFIE [159], InstructDiffusion [160], Emu Edit [161], DialogPaint [162], Inst-Inpaint [163], HIVE [164], ImageBrush [165], InstructAny2Pix [166], MGIE [167], SmartEdit [168] Pseudo-Target Retrieval with Weak Supervision iEdit [169], TDIELR [170], ChatFace [171] Fig. 2: Taxonomy of training-based approaches for image editing.…”
mentioning
confidence: 99%