Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-tutorials.3
|View full text |Cite
|
Sign up to set email alerts
|

Meta Learning and Its Applications to Natural Language Processing

Abstract: Deep learning has been the mainstream technique in natural language processing (NLP) area. However, the techniques require many labeled data and are less generalizable across domains. Meta-learning is an arising field in machine learning studying approaches to learn better learning algorithms. Approaches aim at improving algorithms in various aspects, including data efficiency and generalizability. Efficacy of approaches has been shown in many NLP tasks, but there is no systematic survey of these approaches in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 92 publications
0
4
0
Order By: Relevance
“…Meta-Learning for Semantic Parsing A variety of NLP applications have adopted metalearning in zero-and few-shot learning scenarios as a method of explicitly training for generalization (Lee et al, 2021;Hedderich et al, 2021). Within semantic parsing, there has been increasing interest in cross-database generalization, motivated by datasets such as Spider (Yu et al, 2018) requiring navigation of unseen databases (Herzig and Berant, 2017;Suhr et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Meta-Learning for Semantic Parsing A variety of NLP applications have adopted metalearning in zero-and few-shot learning scenarios as a method of explicitly training for generalization (Lee et al, 2021;Hedderich et al, 2021). Within semantic parsing, there has been increasing interest in cross-database generalization, motivated by datasets such as Spider (Yu et al, 2018) requiring navigation of unseen databases (Herzig and Berant, 2017;Suhr et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…built between all users, and then, it is personalized for each client using their data (Kulkarni et al, 2020;Schneider and Vlachos, 2019;Lee et al, 2021). In such cases, each user has either an entirely separate model, or additional personal parameters, causing significant overheads, both in terms of storage of the large models, and the computation complexity of training separate models for each user.…”
Section: Random User Identifiermentioning
confidence: 99%
“…English + Indic Train: This approach combines approaches (1) and (2). The model is first pre-finetuned (Lee et al, 2021;Aghajanyan et al, 2021) on English XNLI data and then finetuned on Indic language of IN-DICXNLI data.…”
Section: Training-evaluation Strategiesmentioning
confidence: 99%