2017
DOI: 10.1186/s41039-017-0051-y
|View full text |Cite
|
Sign up to set email alerts
|

Evaluation of automatically generated English vocabulary questions

Abstract: This paper describes details of the evaluation experiments for questions created by an automatic question generation system. Given a target word and one of its word senses, the system generates a multiple-choice English vocabulary question asking for the closest in meaning to the target word in the reading passage. Two kinds of evaluation were conducted considering two aspects: (1) measuring English learners' proficiency and (2) their similarity to the human-made questions. The first evaluation is based on the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 15 publications
(10 citation statements)
references
References 13 publications
0
10
0
Order By: Relevance
“…Different qualitative and quantitative analyses were carried out to evaluate questions auto-generated from domain ontologies (Alsubait et al 2014;Vinu and Kumar 2017;Seyler et al 2016;Susanti et al 2017). Papasalouros et al (2017;2011) autogenerated multiple choice questions (MCQs) from the Eupalineio Tunnel ontology, which is a domain ontology about the ancient Greek history.…”
Section: Related Workmentioning
confidence: 99%
“…Different qualitative and quantitative analyses were carried out to evaluate questions auto-generated from domain ontologies (Alsubait et al 2014;Vinu and Kumar 2017;Seyler et al 2016;Susanti et al 2017). Papasalouros et al (2017;2011) autogenerated multiple choice questions (MCQs) from the Eupalineio Tunnel ontology, which is a domain ontology about the ancient Greek history.…”
Section: Related Workmentioning
confidence: 99%
“…Note that the lower perplexity indicates the generated questions are more likely to human writing. Besides, we also use human evaluation that is frequently adopted in recent works about question generation [5,22,10]. In human evaluation, all generated questions mixed together are evaluated by human, and each question is rated by two people.…”
Section: Methodsmentioning
confidence: 99%
“…Unlike this previous research, the aim of the present study is to control item difficulty in the automatic question generation task. Automatic question generation has been proven to be able to generate questions with a quality that is comparable to those made by humans (Susanti et al 2017). Moreover, they cannot be distinguished from human-made questions (Le and Pinkwart 2015; Susanti et al 2017).…”
Section: Introductionmentioning
confidence: 99%