2019
DOI: 10.1109/tip.2019.2910052
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Level Semantic Feature Augmentation for One-Shot Learning

Abstract: The ability to quickly recognize and learn new visual concepts from limited samples enables humans to quickly adapt to new tasks and environments. This ability is enabled by semantic association of novel concepts with those that have already been learned and stored in memory. Computers can start to ascertain similar abilities by utilizing a semantic concept space. A concept space is a high-dimensional semantic space in which similar abstract concepts appear close and dissimilar ones far apart. In this paper, w… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
109
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 243 publications
(109 citation statements)
references
References 60 publications
(103 reference statements)
0
109
0
Order By: Relevance
“…[8], CP-ANN [6], PMN, and PMN w/ H [30]. (3) The data augmentation methods are also compared -flipping: the input image is flipped from left to right; Gaussian noise: cross-validated Gaussian noise N (0, 10) is added to each pixel of the input image; Gaussian noise (feature level): cross-validated Gaussian noise N (0, 0.3) is added to each dimension of the ResNet feature for each image; Mixup: using mixup [35] to combine probe and gallery images. For fair comparisons, all these augmentation methods use the prototype classifier as the one-shot classifier.…”
Section: Results On Imagenet 1k Challengementioning
confidence: 99%
See 1 more Smart Citation
“…[8], CP-ANN [6], PMN, and PMN w/ H [30]. (3) The data augmentation methods are also compared -flipping: the input image is flipped from left to right; Gaussian noise: cross-validated Gaussian noise N (0, 10) is added to each pixel of the input image; Gaussian noise (feature level): cross-validated Gaussian noise N (0, 0.3) is added to each dimension of the ResNet feature for each image; Mixup: using mixup [35] to combine probe and gallery images. For fair comparisons, all these augmentation methods use the prototype classifier as the one-shot classifier.…”
Section: Results On Imagenet 1k Challengementioning
confidence: 99%
“…However, the generated images in this way are particularly subject to visual similarity with the original images. In addition to adding noise or jittering, previous work seeks to augment training images by using semi-supervised techniques [31,18,16] and utilizing relation between visual and semantic representations [3] , or directly synthesizing new instances in the feature domain [9,30,21,6] to transfer knowledge of data distribution from base classes to novel classes. By contrast, we also use samples from base classes to help synthesize deformed images but directly aim at maximizing the one-shot recognition accuracy.…”
Section: Related Workmentioning
confidence: 99%
“…To address this critical problem, a classifier should be implemented to distinguish the new categories. To the best of our knowledge, current researches commonly choose one of the following classifiers to gain their best performance, such as SVM [53], cosine-similarity [47] and nearest neighbor.…”
Section: ) Convnet-based Feature Extractormentioning
confidence: 99%
“…However, the miniImageNet consists of 60,000 color images with 100 classes of which 64 classes for training. The training data is enough to learn a good feature extractor for a common few-shot classification task, and nearly 80% accuracy has been already achieved recently [53]. In this paper, we focus on the fine-grained few-shot classification tasks.…”
Section: A Experimental Designmentioning
confidence: 99%
“…Meta Network proposed by Munkhdalai et al [40] allows fast generalization through shift of inductive bias using meta learner. Besides, lots of researchers try to generate auxiliary samples using a few samples in a class to alleviate overfitting [41][42][43] where Generative Adversarial Network (GAN) is commonly used. Generating process called hallucination is often guided by attribute, semantics and regularization to reduce deviation between results and original samples.…”
Section: Few-shot Learningmentioning
confidence: 99%