2022
DOI: 10.1007/978-3-031-19821-2_3
|View full text |Cite
|
Sign up to set email alerts
|

Revisiting the Critical Factors of Augmentation-Invariant Representation Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 23 publications
0
2
0
Order By: Relevance
“…It has been a notorious problem that fine-tuning LARStrained models with optimization hyper-parameters best selected for SGD-trained counterparts yields sub-optimal performance [11,31]. To address this, recent work [31] proposes NormRescale to scale the norm of LARS-trained weights by a specific anchor (e.g., the norm of SGD-trained weights or a constant number). It helps the LARS-trained models fit to the optimization strategy of fine-tuning.…”
Section: Evaluation Setupmentioning
confidence: 99%
See 1 more Smart Citation
“…It has been a notorious problem that fine-tuning LARStrained models with optimization hyper-parameters best selected for SGD-trained counterparts yields sub-optimal performance [11,31]. To address this, recent work [31] proposes NormRescale to scale the norm of LARS-trained weights by a specific anchor (e.g., the norm of SGD-trained weights or a constant number). It helps the LARS-trained models fit to the optimization strategy of fine-tuning.…”
Section: Evaluation Setupmentioning
confidence: 99%
“…Most of checkpoints are from the official repository, except for BYOL [21]. The ResNet-50 [25] pre-trained by BYOL are from the implementation of [31]. We list out the URLs for downloading these models:…”
Section: A3 Distilling From Different Teachersmentioning
confidence: 99%