2021
DOI: 10.1007/978-3-030-92307-5_76
|View full text |Cite
|
Sign up to set email alerts
|

TRGAN: Text to Image Generation Through Optimizing Initial Image

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 10 publications
0
5
0
Order By: Relevance
“…DDPO has therefore been used to learn how to generate images that are more compressed or uncompressed, by minimizing or maximizing JPEG compression; more aesthetically pleasing, by maximizing LAION score (Schuhmann, 2022); or more prompt-aligned, by maximizing the similarity between the embeddings of prompt and generated image description. Improving the aesthetics of the image while preserving the text-image alignment has also been done at the prompt level (Hao et al, 2023). A language model that given human input provides an optimized prompt can be trained with PPO to maximize both an aesthetic score (from an aesthetic predictor) and a relevance score (as CLIP embedding similarity) of the image generated from the given prompt.…”
Section: Overviewmentioning
confidence: 99%
“…DDPO has therefore been used to learn how to generate images that are more compressed or uncompressed, by minimizing or maximizing JPEG compression; more aesthetically pleasing, by maximizing LAION score (Schuhmann, 2022); or more prompt-aligned, by maximizing the similarity between the embeddings of prompt and generated image description. Improving the aesthetics of the image while preserving the text-image alignment has also been done at the prompt level (Hao et al, 2023). A language model that given human input provides an optimized prompt can be trained with PPO to maximize both an aesthetic score (from an aesthetic predictor) and a relevance score (as CLIP embedding similarity) of the image generated from the given prompt.…”
Section: Overviewmentioning
confidence: 99%
“…While manual prompt engineering can lead to significant progress, the process of designing prompts requires time and experience, and may not always yield optimal results. Therefore, various methods have focused on automatically searching for prompts through mining (Jiang et al 2020), parsing (Haviv, Berant, and Globerson 2021), text generation (Hao et al 2023) and LLMs (Zhu et al 2023;Chakrabarty et al 2023). Additionally, many previous works have focused on gradient-based prompt learning methods, such as (Zhou et al 2022;Wen et al 2023).…”
Section: Prompt Engineeringmentioning
confidence: 99%
“…Our work is closely related to prior research in the field of prompt engineering. For example, Google's recent study (Hao et al 2023) introduced a reinforcement learningbased approach to prompt training. However, their strategy is essentially a training methodology that can be applied to other models.…”
Section: Prompt Engineeringmentioning
confidence: 99%
See 1 more Smart Citation
“…Previous studies (Peng et al, 2023b Optimizing Prompts. Previous studies have explored various methods to optimize prompts, such as tuning soft prompts (Qin and Eisner, 2021;Liu et al, 2023a) or training auxiliary models (Hao et al, 2022;Zhang et al, 2023b). To address the need for extensive model training, the gradientfree prompting technique CoT (Kojima et al, 2022;Zhang et al, 2022; has been proposed to enhance reasoning abilities.…”
Section: Related Workmentioning
confidence: 99%