2022
DOI: 10.1007/978-3-031-19836-6_5
|View full text |Cite
|
Sign up to set email alerts
|

StoryDALL-E: Adapting Pretrained Text-to-Image Transformers for Story Continuation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
27
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 30 publications
(27 citation statements)
references
References 31 publications
0
27
0
Order By: Relevance
“…To address challenges in human evaluation due to ad-hoc practices, prior studies offer shared human evaluation protocols, such as the one for unconditional image synthesis [39]. NLG community provides in-depth analysis of challenges in human evaluation on story generation [16]. To facilitate reliable model comparison based on human evaluation, a platform hosting human evaluation of multiple language generation tasks is proposed [18].…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…To address challenges in human evaluation due to ad-hoc practices, prior studies offer shared human evaluation protocols, such as the one for unconditional image synthesis [39]. NLG community provides in-depth analysis of challenges in human evaluation on story generation [16]. To facilitate reliable model comparison based on human evaluation, a platform hosting human evaluation of multiple language generation tasks is proposed [18].…”
Section: Related Workmentioning
confidence: 99%
“…Human evaluation is typically done in a crowdsourcing platform, and there are many well-known practices. Yet, quality control is an open challenge [16]. Recently, a platform for benchmarking multiple NLG tasks was launched [18].…”
Section: Lack Of Quality Checkmentioning
confidence: 99%
See 2 more Smart Citations
“…Another pioneering work in the field of story synthesis is StoryGAN [24], a conditional GAN [29] trained to generate visual sequences corresponding to text descriptions of a story. Subsequently, StoryDALL-E [28] proposed the adaptation of pre-trained text-to-image transformers to generate visual sequences that continue a given text-based story, introducing a new task of story continuation. More recently, Pan et al [32] proposed history-aware Auto-Regressive Latent Diffusion Models that leverage diffusion models for story synthesis, utilizing a history-aware conditioning network to encode past caption-image pairs.…”
Section: Introductionmentioning
confidence: 99%