2021
DOI: 10.1109/tpami.2021.3068449
|View full text |Cite
|
Sign up to set email alerts
|

Ranked List Loss for Deep Metric Learning

Abstract: The objective of deep metric learning (DML) is to learn embeddings that can capture semantic similarity and dissimilarity information among data points. Existing pairwise or tripletwise loss functions used in DML are known to suffer from slow convergence due to a large proportion of trivial pairs or triplets as the model improves. To improve this, ranking-motivated structured losses are proposed recently to incorporate multiple examples and exploit the structured information among them. They converge faster an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(12 citation statements)
references
References 53 publications
(138 reference statements)
0
12
0
Order By: Relevance
“…3. List-based: these loss functions also make use of distances between pairs of images but, instead of computing the loss directly over them (as the pair-based do), distances are used to build intermediate list-like structures, such as soft-binning histograms, from which a final loss is derived [112,113,114,115]. Since these losses take into account (potentially) long lists containing all images in the training set, they circumvent the problem of locality, and provide a global analysis of the feature space.…”
Section: Cnn Loss Functions To Learn Image Global Representations For Cbirmentioning
confidence: 99%
“…3. List-based: these loss functions also make use of distances between pairs of images but, instead of computing the loss directly over them (as the pair-based do), distances are used to build intermediate list-like structures, such as soft-binning histograms, from which a final loss is derived [112,113,114,115]. Since these losses take into account (potentially) long lists containing all images in the training set, they circumvent the problem of locality, and provide a global analysis of the feature space.…”
Section: Cnn Loss Functions To Learn Image Global Representations For Cbirmentioning
confidence: 99%
“…Multi-level embeddings, i.e. the generation of embedding by pooling information at different stages of the CNN, have also been used for extracting more robust feature representations in metric learning methods [12], [13]. However, in the vehicle ReID field, they have not been fully explored, despite the apparent benefit of multi-level embeddings to the problem.…”
Section: Related Workmentioning
confidence: 99%
“…Given triplets of samples projected to an embedding space, the purpose of triplet loss is to make the distance between the anchor and the negative point larger than the distance between the anchor and the positive point. Recently, group-based metric learning loss functions have been proposed [13], [18], [21], and achieve superior results in retrieval problems, since they incorporate more than three samples, and take into consideration valuable information among multiple data points. In several works [8], [14], classification loss is also used as a complement to similarity metric objectives, in order to increase the discriminative power of the network.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Most contributions in deep metric learning have been devoted to the design of loss functions. The majority of them belong to pair-based functions, e.g., contrastive loss [6,12], triplet loss [7,13], N-pair loss [14], hierarchical triplet loss [15], ranked list loss [16], and multi-similarity loss with general pair weighting [17]. Besides, some loss functions adopt proxy mechanism, such as proxy Neighborhood Component Analysis (ProxyNCA) [18] and proxy anchor [19], to speedup the convergence of model training, where the optimization is carried out on the proxies of triplets, each of which involves an anchor point and similar/dissimilar proxy points.…”
Section: Introductionmentioning
confidence: 99%