Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.448
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Few-Shot Learn Across Diverse Natural Language Classification Tasks

Abstract: Pre-trained transformer models have shown enormous success in improving performance on several downstream tasks. However, fine-tuning on a new task still requires large amounts of taskspecific labeled data to achieve good performance. We consider this problem of learning to generalize to new tasks with a few examples as a meta-learning problem. While meta-learning has shown tremendous progress in recent years, its application is still limited to simulated problems or problems with limited diversity across task… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
99
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 72 publications
(101 citation statements)
references
References 45 publications
1
99
1
Order By: Relevance
“…Pre-trained language models have also been applied to few-shot text classification. LEOPARD (Bansal et al, 2020) uses BERT (Devlin et al, 2019) with optimization-based metalearning framework to achieve good performance on diverse NLP classification tasks. More recently, GPT-3 (Brown et al, 2020) shows that the language model itself can be used to perform few-shot text classification without using meta-learning.…”
Section: Related Workmentioning
confidence: 99%
“…Pre-trained language models have also been applied to few-shot text classification. LEOPARD (Bansal et al, 2020) uses BERT (Devlin et al, 2019) with optimization-based metalearning framework to achieve good performance on diverse NLP classification tasks. More recently, GPT-3 (Brown et al, 2020) shows that the language model itself can be used to perform few-shot text classification without using meta-learning.…”
Section: Related Workmentioning
confidence: 99%
“…Since these applications come with well-defined task distributions, they do not have the same overfitting challenges. On the other hand, many works deal with few-shot adaptation in settings with no clear task distribution (Dou et al, 2019;Bansal et al, 2020a) but do not address meta-overfitting, and thus are complementary to our work.…”
Section: Related Workmentioning
confidence: 99%
“…On GLUE-SciTail, we compare against SMLMT (Bansal et al, 2020b) and find that MAML-DRECA improves over it by 1.5 accuracy points. However, we note that the confidence intervals of these approaches overlap, and also that (Bansal et al, 2020a) consider the entire GLUE data to train the meta-learner whereas we only consider NLI datasets within GLUE. Table 3: Results on NLI few-shot learning.…”
Section: Modelsmentioning
confidence: 99%
“…Text classification has a vast spectrum of applications, such as sentiment classification and intent classification. The meta-learning algorithms developed for image classification can be applied to text classification with slight modification to incorporate domain knowledge in each application Tan et al, 2019;Geng et al, 2019;Dou et al, 2019;Bansal et al, 2019).…”
Section: Text Classificationmentioning
confidence: 99%