Proceedings of the 24th International Conference on Machine Learning 2007
DOI: 10.1145/1273496.1273523
|View full text |Cite
|
Sign up to set email alerts
|

Information-theoretic metric learning

Abstract: We formulate the metric learning problem as that of minimizing the differential relative entropy between two multivariate Gaussians under constraints on the Mahalanobis distance function. Via a surprising equivalence, we show that this problem can be solved as a low-rank kernel learning problem. Specifically, we minimize the Burg divergence of a low-rank kernel to an input kernel, subject to pairwise distance constraints. Our approach has several advantages over existing methods. First, we present a natural in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
1,504
1
6

Year Published

2011
2011
2020
2020

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 1,637 publications
(1,512 citation statements)
references
References 9 publications
1
1,504
1
6
Order By: Relevance
“…We use the default settings for c and l in the authors' code [12]. The setting of K determines "how local" the learner is; its optimal setting depends on the training data and query.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We use the default settings for c and l in the authors' code [12]. The setting of K determines "how local" the learner is; its optimal setting depends on the training data and query.…”
Section: Methodsmentioning
confidence: 99%
“…We employ the information-theoretic metric learning (ITML) algorithm [12], due to its efficiency and kernelizability. Figure 4: Example fine-grained neighbor pairs for three test pairs (top row) from the datasets tested in this paper.…”
Section: Selecting Fine-grained Neighboring Pairsmentioning
confidence: 99%
“…with x and x ′ image indexes and A 0 a positive semi-definite matrix, that can be learnt using optimization [43,11,24] or boosting [17]. The method in [11] is interesting for our context, since the algorithm seems to be able to update matrix A for each new label.…”
Section: Kernel Learningmentioning
confidence: 99%
“…The method in [11] is interesting for our context, since the algorithm seems to be able to update matrix A for each new label. All values of matrix A are subject to change, and thus all distance values between labelled and labelled/unlabelled images must be recomputed.…”
Section: Kernel Learningmentioning
confidence: 99%
“…Formally, the Mahanalobis distance between two data x, y ∈ R d is MD(x, y) = (x − y) T M(x − y), the goal is to learn a proper symmetric positive definite matrix M ∈ R d×d . Information theoretic metric learning (ITML) [6] is one of the state-of-the-art methods for Mahalanobis metric learning which uses an information theoretic approach to optimize M under the constraints that the similarity between each pair labeled "same" is below a specified threshold and the one between each pair labeled "different" is above another specified threshold. Chechik et al [42] learnt a parametric similarity function which gives supervision on the relative similarity between two pairs of images through a bilinear form.…”
Section: Introductionmentioning
confidence: 99%