Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 2021
DOI: 10.18653/v1/2021.findings-acl.132
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Generate Questions by Learning to Recover Answer-containing Sentences

Abstract: To train a question answering model based on machine reading comprehension (MRC), significant effort is required to prepare annotated training data composed of questions and their answers from contexts. Recent research has focused on synthetically generating a question from a given context and an annotated (or generated) answer by training an additional generative model to augment the training data. In light of this research direction, we propose a novel pre-training approach that learns to generate contextual… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 13 publications
(12 citation statements)
references
References 24 publications
0
12
0
Order By: Relevance
“…2019), more attention has been paid to this issue (Back et al . 2020; Liu et al . 2020a; Zhang et al .…”
Section: Future Trends and Opportunitiesmentioning
confidence: 99%
“…2019), more attention has been paid to this issue (Back et al . 2020; Liu et al . 2020a; Zhang et al .…”
Section: Future Trends and Opportunitiesmentioning
confidence: 99%
“…However, their human evaluation centered around question correctness and fluency rather than relevance of answer selection. Similar follow-up studies also fail to explicitly ask annotators whether or not the extracted answers, and subsequent generated questions, were relevant to the broader topic of the context passage (Willis et al, 2019;Cui et al, 2021;Du and Cardie, 2018;Alberti et al, 2019;Back et al, 2021).…”
Section: Related Work and Backgroundmentioning
confidence: 99%
“…We also compare our pre-training to that of the data augmentation (Mao et al, 2021a;Oguz et al, 2021). We generate 'questionanswer' pairs from a Wikipedia dump using a question and answer generation model trained on the NQ dataset using ASGen approach (Back et al, 2021). The model is further trained after the pretraining by this augmented data for 15,000 steps before fine-tuning.…”
Section: Effectiveness Of Emss Pre-trainingmentioning
confidence: 99%