2022
DOI: 10.48550/arxiv.2210.09236
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ZooD: Exploiting Model Zoo for Out-of-Distribution Generalization

Abstract: Recent advances on large-scale pre-training have shown great potentials of leveraging a large set of Pre-Trained Models (PTMs) for improving Out-of-Distribution (OoD) generalization, for which the goal is to perform well on possible unseen domains after fine-tuning on multiple training domains. However, maximally exploiting a zoo of PTMs is challenging since fine-tuning all possible combinations of PTMs is computationally prohibitive while accurate selection of PTMs requires tackling the possible data distribu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
11
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(11 citation statements)
references
References 43 publications
0
11
0
Order By: Relevance
“…PTMs for domain generalization. Methods leveraging pretraining models have shown promising improvements in domain generalization performance (Wiles et al, 2022;Arpit et al, 2021;Dong et al, 2022;Wortsman et al, 2022;Rame et al, 2022;Ramé et al, 2022). Among them, ensemble methods combined with PTMs show further advantages.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…PTMs for domain generalization. Methods leveraging pretraining models have shown promising improvements in domain generalization performance (Wiles et al, 2022;Arpit et al, 2021;Dong et al, 2022;Wortsman et al, 2022;Rame et al, 2022;Ramé et al, 2022). Among them, ensemble methods combined with PTMs show further advantages.…”
Section: Related Workmentioning
confidence: 99%
“…Arpit et al (2021) ensemble the predictions of moving average models. Recent methods Dong et al, 2022) further consider the ensemble of models with different architectures to exploit the growing large PTM hubs. Specifically, ensemble predictions of multiple different PTMs via instance-specific attention weights.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations