2022
DOI: 10.48550/arxiv.2201.03529
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Head2Toe: Utilizing Intermediate Representations for Better Transfer Learning

Abstract: Transfer-learning methods aim to improve performance in a data-scarce target domain using a model pretrained on a data-rich source domain. A cost-efficient strategy, linear probing, involves freezing the source model and training a new classification head for the target domain. This strategy is outperformed by a more costly but state-of-the-art method-fine-tuning all parameters of the source model to the target domain-possibly because fine-tuning allows the model to leverage useful information from intermediat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

1
0
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 23 publications
1
0
0
Order By: Relevance
“…8 we show the improvement achieved through ensembling the final output with each of the intermediate layers over using the final layer only (with the ViT-B/16 encoder). Similar to Evci et al (2022), we show that there always exists a hidden layer ensembling which always leads to a slight improvement of the CL accuracy (e.g. 2.5%-points on FGVCAircraft).…”
Section: Data Analysis: Pre-training and Downstreamsupporting
confidence: 69%
“…8 we show the improvement achieved through ensembling the final output with each of the intermediate layers over using the final layer only (with the ViT-B/16 encoder). Similar to Evci et al (2022), we show that there always exists a hidden layer ensembling which always leads to a slight improvement of the CL accuracy (e.g. 2.5%-points on FGVCAircraft).…”
Section: Data Analysis: Pre-training and Downstreamsupporting
confidence: 69%