Proceedings of the Tenth Workshop on Innovative Use of NLP for Building Educational Applications 2015
DOI: 10.3115/v1/w15-0618
|View full text |Cite
|
Sign up to set email alerts
|

RevUP: Automatic Gap-Fill Question Generation from Educational Texts

Abstract: This paper describes RevUP which deals with automatically generating gap-fill questions. RevUP consists of 3 parts: Sentence Selection, Gap Selection & Multiple Choice Distractor Selection. To select topicallyimportant sentences from texts, we propose a novel sentence ranking method based on topic distributions obtained from topic models. To select gap-phrases from each selected sentence, we collected human annotations, using the Amazon Mechanical Turk, on the relative relevance of candidate gaps. This data is… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
41
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 49 publications
(48 citation statements)
references
References 7 publications
0
41
0
Order By: Relevance
“…Models with recent advances of deep learning techniques are even capable of exceeding human performance in some RC tasks, such as for questions with span-based answers ). However, it is not the case when directly applying the state-of-the-art models to multiple choice questions (MCQs) in RACE dataset (Lai et al 2017), elaborately designed by human experts for real examinations, where the task is to select the correct answer from a few given options after reading the article. The performance gap between the state-of-the-art deep models (53.3%) (Tay, Tuan, and Hui 2018) and ceiling (95%) (Lai et al 2017) is significant.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Models with recent advances of deep learning techniques are even capable of exceeding human performance in some RC tasks, such as for questions with span-based answers ). However, it is not the case when directly applying the state-of-the-art models to multiple choice questions (MCQs) in RACE dataset (Lai et al 2017), elaborately designed by human experts for real examinations, where the task is to select the correct answer from a few given options after reading the article. The performance gap between the state-of-the-art deep models (53.3%) (Tay, Tuan, and Hui 2018) and ceiling (95%) (Lai et al 2017) is significant.…”
Section: Introductionmentioning
confidence: 99%
“…First, a distractor candidate set is extracted from multiple sources, such as GloVe vocabulary (Pennington, Socher, and Manning 2014), noun phrases from textbooks (Welbl, Liu, and Gardner 2017) and articles (Araki et al 2016). Then similarity based (Guo et al 2016;Stasaski and Hearst 2017;Kumar, Banchs, and D'Haro 2015;Mitkov 2003;Zhou, Lyu, and King 2012;Yang et al 2011) or learning based (Liang et al 2018Sakaguchi, Arase, and Komachi 2013;Liang et al 2017;Yang et al 2010) algorithms are employed to select the distractors. Another manner is to apply some pre-defined rules to prepare distractors by changing the surface form of some words or phrases (Chen, Liou, and Chang 2006).…”
Section: Introductionmentioning
confidence: 99%
“…Several earlier works process documents as individual sentences using syntactic (Heilman and Smith, 2010;Ali et al, 2010;Kumar et al, 2015) or semantic-based parsing (Mannem et al, 2010;Lindberg et al, 2013), then reformulate questions using hand-crafted rules acting on parse trees. These traditional approaches generate questions with a high word overlap with the original text that pertain specifically to the given sentence by re-arranging the sentence parse tree.…”
Section: Related Workmentioning
confidence: 99%
“…A good distractor could be a synonym of the key phrase or an important term in the domain of the key phrase. Distractors in Revup, an AQG, are selected from word2vec, a vector of words 16 . Text summarization features like length of a sentence, number of common tokens, number of noun and pronouns, and position of a sentence are generally considered 17 .…”
Section: Gap-fill Questionsmentioning
confidence: 99%