2018
DOI: 10.48550/arxiv.1803.02999
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On First-Order Meta-Learning Algorithms

Abstract: This paper considers meta-learning problems, where there is a distribution of tasks, and we would like to obtain an agent that performs well (i.e., learns quickly) when presented with a previously unseen task sampled from this distribution. We analyze a family of algorithms for learning a parameter initialization that can be fine-tuned quickly on a new task, using only firstorder derivatives for the meta-learning updates. This family includes and generalizes first-order MAML, an approximation to MAML obtained … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
863
1

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 554 publications
(933 citation statements)
references
References 9 publications
3
863
1
Order By: Relevance
“…While firstorder MAML requires less memory and compute for each update, it performed significantly worse than MAML. Other first-order MAML methods such as REPTILE (Nichol et al, 2018) are an alternative that can reduce memory usage at the cost of more compute time and some accuracy.…”
Section: Discussionmentioning
confidence: 99%
“…While firstorder MAML requires less memory and compute for each update, it performed significantly worse than MAML. Other first-order MAML methods such as REPTILE (Nichol et al, 2018) are an alternative that can reduce memory usage at the cost of more compute time and some accuracy.…”
Section: Discussionmentioning
confidence: 99%
“…The meta-model is built so that it can be rapidly adapted (online) to any new learning task that may be encountered, exploiting just a few experiences from the new task. The works in [11], [12] validate a meta-learning framework that can be used in several learning tasks, e.g., it can be applied to both supervised ML (regression and classification) and RL scenarios. Other works propose metalearning for more specific scenarios, i.e., the update rule and selective copy of weights of deep networks [36], [37], [38] and recurrent networks [39], [40], [41].…”
Section: Learning Concepts In Networkingmentioning
confidence: 99%
“…Once derived with the above procedure, meta-model Θ can be used as a starting point for finding any specific model that suits a newly encountered task, by only using a small amount of experience collected on this new task [11], [12]. In our scenario, the learning tasks of FALCON are the different network conditions it may encounter and it should adapt to by deriving specific scheduling policies.…”
Section: A Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…Meta reinforcement learning (e.g., Duan et al, 2016;Finn et al, 2017;Nichol et al, 2018;Xu et al, 2018;Rakelly et al, 2019;Zintgraf et al, 2020) can be seen as a generalized settings of reward transfer, where the difference between the tasks can also be different in the underlying dynamics. And usually they still need few-shot interactions with the environment to generalize, differ from our pure offline settings.…”
Section: Related Workmentioning
confidence: 99%