2017
DOI: 10.48550/arxiv.1704.04683
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

RACE: Large-scale ReAding Comprehension Dataset From Examinations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
133
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 138 publications
(133 citation statements)
references
References 7 publications
0
133
0
Order By: Relevance
“…Unlike the previous three datasets, Nar-rativeQA [35] does not restrict the answers to be the span of texts in the articles, therefore, it can be used as an answer-abstraction QG dataset. Race [38], McTest [64], OpenbookQA [49], and ARC [15] are commonly used multi-choice QG datasets. BoolQA [14] is a typical boolean QG dataset, gathered from Google search engine.…”
Section: Related Work 21 Question Generationmentioning
confidence: 99%
See 1 more Smart Citation
“…Unlike the previous three datasets, Nar-rativeQA [35] does not restrict the answers to be the span of texts in the articles, therefore, it can be used as an answer-abstraction QG dataset. Race [38], McTest [64], OpenbookQA [49], and ARC [15] are commonly used multi-choice QG datasets. BoolQA [14] is a typical boolean QG dataset, gathered from Google search engine.…”
Section: Related Work 21 Question Generationmentioning
confidence: 99%
“…• Answer-extraction QG: SQuADv1.1 [58] • Answer-abstraction QG: NarrativeQA [35] • Multi-choice QG: RACE [38], McTest [64], OpenbookQA [49], ARC-easy, ARC-hard [15] • Boolean QG: BoolQA [14] We assume datasets arrive in the following order: "McTest→SQuAD →RACE→NarrativeQA→Arc-easy→Arc-hard→OpenbookQA →BoolQA", which corresponds to the exact release dates of these datasets in the real world. Details on dataset characteristics, statistics, and splitting strategies are in Appendix A.1.…”
Section: Experiments 51 Datasetsmentioning
confidence: 99%
“…Multiple-choice datasets overcome the evaluation difficulty in abstractive datasets by simply asking the reader to select the correct answer from the candidate options. Representative datasets, such as RACE [19] and DREAM [33], utilize exam questions collected from standard English proficiency tests, which are generated by language teachers to evaluate a variety of language capabilities of non-native English speakers.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, the C 3 dataset [34] was released, which contains both free-form questions and multiple-choice answer options. C 3 is collected using the exam questions for Chinese-as-a-second-language tests and consists of two sub-datasets, C 3 D focusing on normal documents and C 3 M on dialogues, which can be viewed as the Chinese counterparts of RACE [19] and DREAM [33] respectively.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation