2022
DOI: 10.1162/tacl_a_00469
|View full text |Cite
|
Sign up to set email alerts
|

LOT: A Story-Centric Benchmark for Evaluating Chinese Long Text Understanding and Generation

Abstract: Standard multi-task benchmarks are essential for developing pretraining models that can generalize to various downstream tasks. Existing benchmarks for natural language processing (NLP) usually focus only on understanding or generating short texts. However, long text modeling requires many distinct abilities in contrast to short texts, such as the modeling of long-range discourse and commonsense relations, and the coherence and controllability of generation. The lack of standardized benchmarks makes it difficu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(21 citation statements)
references
References 41 publications
0
21
0
Order By: Relevance
“…39,093 examples Synthesis and crowd sourcing. LOT [55] Is a continuation 38,000 examples Expert annotations plausible (Chinese)? of collected texts Located Near [146] Is X located near Y?…”
Section: Collections Of Elementary Mathematical Word Problemsmentioning
confidence: 99%
“…39,093 examples Synthesis and crowd sourcing. LOT [55] Is a continuation 38,000 examples Expert annotations plausible (Chinese)? of collected texts Located Near [146] Is X located near Y?…”
Section: Collections Of Elementary Mathematical Word Problemsmentioning
confidence: 99%
“…We use STORAL-ZH [18], the Chinese part of STORAL [6], as the dataset for target tasks. Furthermore, LongLM-base [21] is selected as our model, that has been pre-trained on 120G Chinese long novels. For TAPT, the language model is further trained on unlabeled STORAL-ZH to equip itself with task-awareness knowledge.…”
Section: Storymentioning
confidence: 99%
“…The story must remain thematically consistent across the complete document as well as keeping creativity. OutGen is an outline-conditioned story generation dataset introduced by Guan et al (2021), which requires generating a coherent long-form story conditioned on an outline of characters and events. The outline is a set of out-of-order phrases.…”
Section: B Datasetsmentioning
confidence: 99%
“…The outline is a set of out-of-order phrases. We use the same data split and evaluation metrics provided by Guan et al (2021) 8 .…”
Section: B Datasetsmentioning
confidence: 99%
See 1 more Smart Citation