2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00286
|View full text |Cite
|
Sign up to set email alerts
|

Hybrid-Attention Based Decoupled Metric Learning for Zero-Shot Image Retrieval

Abstract: In zero-shot image retrieval (ZSIR) task, embedding learning becomes more attractive, however, many methods follow the traditional metric learning idea and omit the problems behind zero-shot settings. In this paper, we first emphasize the importance of learning visual discriminative metric and preventing the partial/selective learning behavior of learner in ZSIR, and then propose the Decoupled Metric Learning (DeML) framework to achieve these individually. Instead of coarsely optimizing an unified metric, we d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
36
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 62 publications
(36 citation statements)
references
References 39 publications
0
36
0
Order By: Relevance
“…To address this challenge, we adopt deep metric learning, which essentially 'learns to compare' image pairs by shrinking intra-class distance while expanding inter-class distance. Several deep metric learning approaches have been proposed and been applied in various applications, including face identification [18], person re-identification [19,20], and image retrieval [21]. Metric learning is known to have a stronger discriminative power since the inter-class distance is directly maximized during training.…”
Section: Deep Metric Learningmentioning
confidence: 99%
“…To address this challenge, we adopt deep metric learning, which essentially 'learns to compare' image pairs by shrinking intra-class distance while expanding inter-class distance. Several deep metric learning approaches have been proposed and been applied in various applications, including face identification [18], person re-identification [19,20], and image retrieval [21]. Metric learning is known to have a stronger discriminative power since the inter-class distance is directly maximized during training.…”
Section: Deep Metric Learningmentioning
confidence: 99%
“…The application of zero-shot in super-resolution, i.e, the ZSSR prescription [23], is among the most widely used models in superresolution and has gained increasing interest recently. In addition, the majority of zero-shot methods that have provided a large number of excellent results in recent years are mainly based on segmentation [39], emotion recognition [40], object detection [41], image retrieval [42][43][44][45], image classification [46][47][48] and intelligent learning in machines or robots [49]. In the ZSSR formalism, LR images are downsampled to generate many lower-resolution images (I = I 0 , I 1 , I 2 , ..., I n ), which serve as the HR supervision information called "HR fathers, " then, each HR father is downscaled by the required scale factor s to obtain the corresponding "LR sons.…”
Section: Zero-shotmentioning
confidence: 99%
“…Attention mechanism, widely studied and utilized in recent years, has been successfully applied in many computer vision tasks, such as image classification [6,12], semantic segmentation [27], and visual reasoning [28]. Inspired by the successful application of attention mechanisms in computer vision tasks, some works [30,31,32,40] attempt to introduce attention mechanisms to improve the discrimination of visual features in the few-shot and zero-shot learning task. A multi-attention network is proposed in [32] to solve one-shot learning problem.…”
Section: Attention Mechanismmentioning
confidence: 99%
“…A stacked semantic-guided attention model is proposed in [30] for fine-grained ZSL, which is capable of discovering discriminative subtle visual differences with the guidance of semantic. In the zero-shot retrieval task, [31] proposed a decoupled metric learning framework to learn discriminative visual attributes.…”
Section: Attention Mechanismmentioning
confidence: 99%