2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00888
|View full text |Cite
|
Sign up to set email alerts
|

Image Deformation Meta-Networks for One-Shot Learning

Abstract: Figure 1. Illustration of a variety of image deformations: ghosted (a, b), stitched (c), montaged (d), and partially occluded (e) images. AbstractHumans can robustly learn novel visual concepts even when images undergo various deformations and lose certain information. Mimicking the same behavior and synthesizing deformed instances of new concepts may help visual recognition systems perform better one-shot learning, i.e., learning concepts from one or few examples. Our key insight is that, while the deformed i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
98
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 218 publications
(115 citation statements)
references
References 18 publications
0
98
0
Order By: Relevance
“…Experiments in the second block investigate the effect of the concentrator and projector, respectively. Removing each component alone results in a performance decrease (cases i, ii, iii) 3 . The accuracy is inferior (-3.93%, 1-shot case) if we remove the network part of the concentrator, implying that its dimension reduction and spatial downsampling is important to the final comparisons.…”
Section: Shallow Network Verificationmentioning
confidence: 99%
“…Experiments in the second block investigate the effect of the concentrator and projector, respectively. Removing each component alone results in a performance decrease (cases i, ii, iii) 3 . The accuracy is inferior (-3.93%, 1-shot case) if we remove the network part of the concentrator, implying that its dimension reduction and spatial downsampling is important to the final comparisons.…”
Section: Shallow Network Verificationmentioning
confidence: 99%
“…At present, the biggest challenge in few-shot learning scenarios is consistently the shortage of training samples, and therefore how to acquire more informative and meaningful labeled data is the core of issue. The hallucination based approaches [40][41][42][43] use a small number of labeled samples to generate more hallucination data, with the goal of achieving a robust and powerful network. Zhang et al [40] pretrained a saliency network to segment foreground and background of available image samples and generated additional data by combining the appropriate foregroundbackground pairs.…”
Section: Related Workmentioning
confidence: 99%
“…The authors utilized these operations to generate more hallucination data for training phase, and thus significantly improved classification performance. Chen et al [42] proposed to relate new concepts to the existing ones in semantic space, and leveraged the relationships to generate new additional samples by interpolating among concepts, further facilitating learning.…”
Section: Related Workmentioning
confidence: 99%
“…Most of these works evaluate their algorithms on some small-scale datasets, e.g., miniImageNet with 64 base categories, 16 validation categories, 20 novel categories. Recently, some works (Hariharan and Girshick 2017;Chen et al 2019c;Hui, Chen, and Chen 2019) switch to a more general and practical setting, where the algorithms aim to recognize hundreds of novel concepts with very limited samples given a set of based categories with sufficient training samples. To address this few-shot learning scenario, (Hariharan and Girshick 2017) learn a transformation function to hallucinate additional samples for novel categories.…”
Section: Related Workmentioning
confidence: 99%