2024
DOI: 10.1109/tmm.2023.3290481
|View full text |Cite
|
Sign up to set email alerts
|

DRAN: Detailed Region-Adaptive Normalization for Conditional Image Synthesis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 47 publications
0
1
0
Order By: Relevance
“…Trained on a vast dataset of 400 million (image, text) pairs, it strives to learn a joint semantical space for both images and texts using contrastive loss. Benefiting from its excellent image/text representation ability, CLIP has found extensive applications in different areas, such as domain adaptation [32], image segmentation [33], [34], image generation and editing [12], [13], [35], [36], [37], [38], [39], [40]. In particular, LAFITE [41] and KNN-Diffusion [42] utilize the CLIP image-text feature space and exploit languagefree text-to-image generation.…”
Section: Vision-language Representationsmentioning
confidence: 99%
“…Trained on a vast dataset of 400 million (image, text) pairs, it strives to learn a joint semantical space for both images and texts using contrastive loss. Benefiting from its excellent image/text representation ability, CLIP has found extensive applications in different areas, such as domain adaptation [32], image segmentation [33], [34], image generation and editing [12], [13], [35], [36], [37], [38], [39], [40]. In particular, LAFITE [41] and KNN-Diffusion [42] utilize the CLIP image-text feature space and exploit languagefree text-to-image generation.…”
Section: Vision-language Representationsmentioning
confidence: 99%