2021
DOI: 10.48550/arxiv.2105.02221
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

How Fine-Tuning Allows for Effective Meta-Learning

Abstract: Representation learning has been widely studied in the context of meta-learning, enabling rapid learning of new tasks through shared representations. Recent works such as MAML have explored using fine-tuning-based metrics, which measure the ease by which fine-tuning can achieve good performance, as proxies for obtaining representations. We present a theoretical framework for analyzing representations derived from a MAML-like algorithm, assuming the available tasks use approximately the same underlying represen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
7
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 23 publications
0
7
0
Order By: Relevance
“…Prior works on transfer learning mainly analyze linear probing (Wu et al, 2020;Tripuraneni et al, 2020;Du et al, 2020). In recent work, (Chua et al, 2021) study regularized fine-tuning in an underparameterized regime where there is a unique global optimum. In contrast, our analysis studies the overparameterized regime (mirroring modern settings of zero train loss) where we need to analyze the trajectory of fine-tuning from the pretrained initialization because there is no unique optimizer of the objective function.…”
Section: Related Work and Discussionmentioning
confidence: 99%
“…Prior works on transfer learning mainly analyze linear probing (Wu et al, 2020;Tripuraneni et al, 2020;Du et al, 2020). In recent work, (Chua et al, 2021) study regularized fine-tuning in an underparameterized regime where there is a unique global optimum. In contrast, our analysis studies the overparameterized regime (mirroring modern settings of zero train loss) where we need to analyze the trajectory of fine-tuning from the pretrained initialization because there is no unique optimizer of the objective function.…”
Section: Related Work and Discussionmentioning
confidence: 99%
“…However, our work focuses on the setting where all tasks share one common low-dimensional representation. [CLL21] shows the benefits of task-specific fine-tuning, which is fundamentally different from our refinement step. Our refinement step aims to reduce the representation dimension with slight information loss and help to improve the sample complexity of subsequent tasks.…”
Section: Related Workmentioning
confidence: 95%
“…This observation encourages the idea that there is some inductive bias relating to the initial weights of the model that might explain the success of fine-tuning. [20] and [21] both show that fine-tuned models generalize well when the representation used by the target task is approximately similar to the one used by the source tasks.…”
Section: Related Workmentioning
confidence: 99%