2022
DOI: 10.48550/arxiv.2201.09699
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

EASY: Ensemble Augmented-Shot Y-shaped Learning: State-Of-The-Art Few-Shot Classification with Simple Ingredients

Abstract: Few-shot learning aims at leveraging knowledge learned by one or more deep learning models, in order to obtain good classification performance on new problems, where only a few labeled samples per class are available. Recent years have seen a fair number of works in the field, introducing methods with numerous ingredients. A frequent problem, though, is the use of suboptimally trained models to extract knowledge, leading to interrogations on whether proposed approaches bring gains compared to using better init… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(18 citation statements)
references
References 27 publications
0
17
0
Order By: Relevance
“…We found ViT performed the best accuracy for both 1-shot and 5-shot learning tasks. Note that other one-shot learning methods in previous study can obtain the same degree of accuracy without using such an excellent backbone DNN like ViT (e.g., 82% [40] and 87% [41] with ResNet-12), and thus DONE is not the best method for accuracy. However, these results suggest that DONE with ViT is already at a level of practical uses.…”
Section: Evaluation Of Dnns In Donementioning
confidence: 86%
See 1 more Smart Citation
“…We found ViT performed the best accuracy for both 1-shot and 5-shot learning tasks. Note that other one-shot learning methods in previous study can obtain the same degree of accuracy without using such an excellent backbone DNN like ViT (e.g., 82% [40] and 87% [41] with ResNet-12), and thus DONE is not the best method for accuracy. However, these results suggest that DONE with ViT is already at a level of practical uses.…”
Section: Evaluation Of Dnns In Donementioning
confidence: 86%
“…Metric learning such as using Siamese network [13] is useful for tasks that require one-shot learning, e.g., face recognition. A Data-augmentation approach generatively increases the number of training inputs [41,18,24]. This approach includes various types such as semi-supervised approaches and example generation using Generative Adversarial Networks [10].…”
Section: Related Workmentioning
confidence: 99%
“…Finally, we conduct comparison of our performance with the stateof-the-art results on three benchmarks: miniImageNet, tieredIm-ageNet, and CUB. We choose the state-of-the-art FSL methods ProtoNet [36], FEAT [47], and ASY [3] as the first group of models.…”
Section: Comparison With the State-of-the-art Methodsmentioning
confidence: 99%
“…There are two main frameworks in the field of FSC: 1) only one unlabelled sample is processed at a time for class predictions, which is called inductive FSC, and 2) the entire unlabelled samples are available for further estimations, which is called transductive FSC. Inductive methods focus on training a feature extractor that generalizes well the embedding in a feature sub-space, they include meta learning methods such as [12,26,2,40,30,37] that train a model in an episodic manner, and transfer learning methods [8,28,48,5,3,33] that train a model with a set of mini-batches.…”
Section: Related Workmentioning
confidence: 99%
“…As for the hyper-parameters, we set T km = 10, T vb = 50, s max = 2 for the balanced setting; T km = 50, T vb = 50, s max = 1 for the unbalanced setting, and we use the same VB priors for all settings. To further show the functionality of our proposed method on different backbones and other benchmarks, we tested BAVARDAGE on a recent high performing feature extractor trained on a ResNet-12 (RN12) neural model [28,3], and we report the accuracy in Table 1 and in Appendix with various settings.…”
Section: Implementation Detailsmentioning
confidence: 99%