2021
DOI: 10.48550/arxiv.2104.02638
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Comparing Transfer and Meta Learning Approaches on a Unified Few-Shot Classification Benchmark

Abstract: Meta and transfer learning are two successful families of approaches to few-shot learning. Despite highly related goals, state-of-the-art advances in each family are measured largely in isolation of each other. As a result of diverging evaluation norms, a direct or thorough comparison of different approaches is challenging. To bridge this gap, we perform a cross-family study of the best transfer and meta learners on both a large-scale meta-learning benchmark (Meta-Dataset, MD), and a transfer learning benchmar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
26
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(26 citation statements)
references
References 28 publications
0
26
0
Order By: Relevance
“…One prominent example is where they show that GPT-3, which is a large transformer model [Vaswani et al, 2017] trained on a large corpus of data, achieves substantial performance on many natural language processing (NLP) tasks and benchmarks in few-shot settings. On image recognition tasks, training on Instagram images [Mahajan et al, 2018] and JFT-300 [Sun et al, 2017] has been proven to be very effective in transfer and few-shot settings [Goyal et al, 2021, Pham et al, 2020, Dosovitskiy et al, 2020, Dumoulin et al, 2021]. Even when no example is provided (zero-shot), CLIP [Radford et al, 2021], which consists of a pair of image encoder and text encoder models trained with a contrastive loss on 400 million image-text pairs from the internet, can achieve remarkable performance.…”
Section: Introductionmentioning
confidence: 99%
“…One prominent example is where they show that GPT-3, which is a large transformer model [Vaswani et al, 2017] trained on a large corpus of data, achieves substantial performance on many natural language processing (NLP) tasks and benchmarks in few-shot settings. On image recognition tasks, training on Instagram images [Mahajan et al, 2018] and JFT-300 [Sun et al, 2017] has been proven to be very effective in transfer and few-shot settings [Goyal et al, 2021, Pham et al, 2020, Dosovitskiy et al, 2020, Dumoulin et al, 2021]. Even when no example is provided (zero-shot), CLIP [Radford et al, 2021], which consists of a pair of image encoder and text encoder models trained with a contrastive loss on 400 million image-text pairs from the internet, can achieve remarkable performance.…”
Section: Introductionmentioning
confidence: 99%
“…If N τ is large (e.g. the recent VTAB+MD benchmark [11] requires a task's support set to be as large as 1000 images), memory on a single GPU is thus quickly exceeded for large images.…”
Section: Our Contributionsmentioning
confidence: 99%
“…query example. In some cases, this can be as many as 1000 images [11]. As a result, the amount of memory required for the computational graph grows linearly with the number of support images, and quadratically with their dimension.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations