2008
DOI: 10.1109/tpami.2007.1140
|View full text |Cite
|
Sign up to set email alerts
|

BoostMap: An Embedding Method for Efficient Nearest Neighbor Retrieval

Abstract: Abstract-This paper describes BoostMap, a method for efficient nearest neighbor retrieval under computationally expensive distance measures. Database and query objects are embedded into a vector space in which distances can be measured efficiently. Each embedding is treated as a classifier that predicts for any three objects X, A, B whether X is closer to A or to B. It is shown that a linear combination of such embedding-based classifiers naturally corresponds to an embedding and a distance measure. Based on t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
70
0

Year Published

2008
2008
2015
2015

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 77 publications
(70 citation statements)
references
References 57 publications
0
70
0
Order By: Relevance
“…Most related to some of the techniques here, Athitsos et al [2,3] propose a boosting-based approach which gives a parametric function for mapping points to binary vectors, and can accommodate metric and non-metric target similarity functions. Salakhutdinov and Hinton [56] use a neural network trained with an NCA objective [26] to build codes for textdocuments.…”
Section: Other Unsupervised Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Most related to some of the techniques here, Athitsos et al [2,3] propose a boosting-based approach which gives a parametric function for mapping points to binary vectors, and can accommodate metric and non-metric target similarity functions. Salakhutdinov and Hinton [56] use a neural network trained with an NCA objective [26] to build codes for textdocuments.…”
Section: Other Unsupervised Methodsmentioning
confidence: 99%
“…Memory usage with LSH is typically greater, however, assuming one opts to mitigate the 0-threshold Hamming distance by expanding the search to multiple independently generated hash tables. 3 Furthermore, whereas a user of Semantic Hashing specifies a radius of interest in the embedded Hamming space, a user of LSH (for the radius-based search variant) specifies the radius of interest in the original feature space.…”
Section: Recap Of Search Strategy Tradeoffsmentioning
confidence: 99%
“…Among them, we focus on pseudo-score based indexing schemes [7]- [10], and use the standard pivot-based indexing scheme [7] and the permutation-based indexing scheme [7], [8] in our experiments in Sect. 5.…”
Section: Metric Space Indexingmentioning
confidence: 99%
“…As there is no external conditions or parameters of the dataset used, we directly used the values reported in the BoostMap paper [3] for other algorithms namely RRO, RLP, FastMap, VP-Trees 1 . Each subplot shows the exact number of DTW distances that needs to be computed against different values of nearest neighbors to be retrieved, for different accuracies on input dataset.…”
Section: Unipen Handwriting Databasementioning
confidence: 99%
“…Although these methods assume that triangular inequality holds, they work for non-metric distances as well with certain amount of distortion in embedding. Athitsos [3] framed embedding construction as a machine learning task, where AdaBoost is used to combine many simple, 1D embeddings into a multidimensional embedding that preserves a significant amount of the proximity structure in original space. These five techniques are most related to our approach as their main target is to learn embeddings for fast retrieval of nearest neighbors.…”
Section: Introductionmentioning
confidence: 99%