2023
DOI: 10.48550/arxiv.2302.12228
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Encoder-based Domain Tuning for Fast Personalization of Text-to-Image Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
16
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(16 citation statements)
references
References 46 publications
0
16
0
Order By: Relevance
“…Fine-tuning generative models for personalization Recent works have focused on customizing and personalizing text-to-image diffusion models by fine-tuning the text embedding [17], full weights [52], or cross-attention layers [34] using a few personalized images. Other works have also investigated training-free approaches for fast adaptation [12,18,30,73]. The idea of fine-tuning only the singular values of weight matrices was introduced by FS-GAN [50] in the GAN literature and further advanced by NaviGAN [14] with an unsupervised method for discovering semantic directions in this compact parameter space.…”
Section: Related Workmentioning
confidence: 99%
“…Fine-tuning generative models for personalization Recent works have focused on customizing and personalizing text-to-image diffusion models by fine-tuning the text embedding [17], full weights [52], or cross-attention layers [34] using a few personalized images. Other works have also investigated training-free approaches for fast adaptation [12,18,30,73]. The idea of fine-tuning only the singular values of weight matrices was introduced by FS-GAN [50] in the GAN literature and further advanced by NaviGAN [14] with an unsupervised method for discovering semantic directions in this compact parameter space.…”
Section: Related Workmentioning
confidence: 99%
“…Fine-tuning aims to adapt the weights of a pretrained generative model to new domains [72,46,71,41,78,34,47,79,30,35,24,44], downstream tasks [70,54,77], and test images [6,53,48,31,25,49]. Several recent works also explore fine-tuning text-toimage models to learn personalized or unseen concepts [33,17,55,18] given a few exemplar images. Similarly, model editing [5,69,19,68,45,38,40,39] aims to modify specific model weights based on users' instructions to incorporate new computational rules or new visual effects.…”
Section: Related Workmentioning
confidence: 99%
“…Note how styles of different fonts are preserved by the semantic modification. , 2023Ruiz et al 2022], and explainability [Chefer et al 2021]. Despite being trained on raster images, their strong visual and semantic priors have also been shown to be successfully applied to other domains, such as motion [Tevet et al 2022], meshes [Michel et al 2021], point cloud [Zhang et al 2021], and vector graphics.…”
Section: Related Workmentioning
confidence: 99%