2022
DOI: 10.1007/978-3-031-20059-5_34
|View full text |Cite
|
Sign up to set email alerts
|

TISE: Bag of Metrics for Text-to-Image Synthesis Evaluation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(2 citation statements)
references
References 24 publications
0
2
0
Order By: Relevance
“…To evaluate individual generated images based on a prompt, previous studies [49,64] often employ metrics based on Contrastive Language-Image Pre-Training (CLIP) [41], which compute text-image consistency by cosine similarity between text and image embeddings in the joint representation space. To better align with human preferences, researchers explored fine-tuned CLIP using datasets of human ratings on images created from identical prompts [16,61,62]. They further utilized scores predicted by the fine-tuned CLIP to approximate human assessment.…”
Section: Evaluation Of Text-to-image Generationmentioning
confidence: 99%
“…To evaluate individual generated images based on a prompt, previous studies [49,64] often employ metrics based on Contrastive Language-Image Pre-Training (CLIP) [41], which compute text-image consistency by cosine similarity between text and image embeddings in the joint representation space. To better align with human preferences, researchers explored fine-tuned CLIP using datasets of human ratings on images created from identical prompts [16,61,62]. They further utilized scores predicted by the fine-tuned CLIP to approximate human assessment.…”
Section: Evaluation Of Text-to-image Generationmentioning
confidence: 99%
“…Evaluation Metrics. Following ) and (Dinh, Nguyen, and Hua 2022), we quantitatively assess the image quality and aesthetic, using the non-reference metrics. Specifically, we choose NIMA (Talebi and Milanfar 2018), MUSIQ (Ke et al 2021), DB-CNN (Zhang et al 2020), and TReS (Golestaneh, Dadsetan, and Kitani 2022).…”
Section: Quantitative Comparisonmentioning
confidence: 99%