Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.491
|View full text |Cite
|
Sign up to set email alerts
|

FewshotQA: A simple framework for few-shot learning of question answering tasks using pre-trained text-to-text models

Abstract: The task of learning from only a few examples (called a few-shot setting) is of key importance and relevance to a real-world setting. For question answering (QA), the current state-of-the-art pre-trained models typically need fine-tuning on tens of thousands of examples to obtain good results. Their performance degrades significantly in a few-shot setting (< 100 examples). To address this, we propose a simple fine-tuning framework that leverages pre-trained text-to-text models and is directly aligned with thei… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 23 publications
(11 citation statements)
references
References 17 publications
0
11
0
Order By: Relevance
“…In particular, we highlight target answer a t as rationale span in the passage c using a special token, "<hl>", following Gu et al (2021). In addition, we adopt a mask prediction scheme that aligns its objective with that of the pre-training phase, shown to be sample efficient in prior work (Chada and Natarajan, 2021).…”
Section: Conversational Answer Extractormentioning
confidence: 99%
See 1 more Smart Citation
“…In particular, we highlight target answer a t as rationale span in the passage c using a special token, "<hl>", following Gu et al (2021). In addition, we adopt a mask prediction scheme that aligns its objective with that of the pre-training phase, shown to be sample efficient in prior work (Chada and Natarajan, 2021).…”
Section: Conversational Answer Extractormentioning
confidence: 99%
“…They are also concatenated with the <sep> token to represent the B. The masked question prediction scheme is inspired by Chada and Natarajan (2021) and we find the scheme is more sample efficient in our preliminary experiment. We train both CQG models for 10 epochs with 16 for batch size, 3e-5 for lr, 0.1 for lr warming up, and 0.01 for weight decay on 2 32GB V100 GPUs.…”
Section: A Implementation Detailsmentioning
confidence: 99%
“…We use the standard encoder-decoder objective of maximizing the log likelihood of text in the ground truth target. For creating input and output prompts, we closely follow the templates proposed by Chada and Natarajan (2021) where prompts are aligned to the format used during the MLM pretraining.…”
Section: Rationale Annotationmentioning
confidence: 99%
“…Since the expensiveness of constructing the annotated datasets, several works have been focused on few-shot learning for KBQA task. Chada et al (2021) proposed a simple fine-tuning framework that regards the query path generation as a text-to-text task [40]. By leveraging a pre-trained sequence-to-sequence models, their method outperforms many state-of-art models with an average margin of 34.2 F1 points on various few-shot settings of multiple QA benchmarks.…”
Section: Related Workmentioning
confidence: 99%