Proceedings of the Forty-Seventh Annual ACM Symposium on Theory of Computing 2015
DOI: 10.1145/2746539.2746553
|View full text |Cite
|
Sign up to set email alerts
|

Optimal Data-Dependent Hashing for Approximate Near Neighbors

Abstract: We show an optimal data-dependent hashing scheme for the approximate near neighbor problem. For an n-point dataset in a d-dimensional space our data structure achieves query time O(d · n ρ+o(1) ) and space O(n 1+ρ+o(1) + d · n), where ρ = 1 2c 2 −1 for the Euclidean space and approximation c > 1. For the Hamming space, we obtain an exponent of ρ = 1 2c−1 . Our result completes the direction set forth in [5] who gave a proof-of-concept that data-dependent hashing can outperform classic Locality Sensitive Hashin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

3
200
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 178 publications
(203 citation statements)
references
References 30 publications
3
200
0
Order By: Relevance
“…Although this work focuses on applying angular LSH to sieving, more generally this work could be considered the first to succeed in applying LSH to lattice algorithms. Various recent followup works have already further investigated the use of different LSH methods [7,8] and other nearest neighbor search methods [9,11,38] in the context of lattice sieving [11][12][13]30,37], and an open problem is whether other lattice algorithms (e.g. provable sieving algorithms, the Voronoi cell algorithm [39]) can benefit from related techniques as well.…”
Section: Introductionmentioning
confidence: 99%
“…Although this work focuses on applying angular LSH to sieving, more generally this work could be considered the first to succeed in applying LSH to lattice algorithms. Various recent followup works have already further investigated the use of different LSH methods [7,8] and other nearest neighbor search methods [9,11,38] in the context of lattice sieving [11][12][13]30,37], and an open problem is whether other lattice algorithms (e.g. provable sieving algorithms, the Voronoi cell algorithm [39]) can benefit from related techniques as well.…”
Section: Introductionmentioning
confidence: 99%
“…Comparing this result to Andoni et al's spherical hash functions h ∈ S [7,8] used in the SphereSieve [33], which have a collision probability of…”
mentioning
confidence: 70%
“…The main difference with previous work [32,33] lies in the choice of the hash function family, which in this paper is the efficient and asymptotically superior cross-polytope LSH, rather than the asymptotically worse angular or hyperplane LSH [13,32] or the less practical spherical LSH [8,33]. This leads to the CPSieve algorithm described in Algorithm 2.…”
Section: Cpsieve: Sieving In Arbitrary Latticesmentioning
confidence: 97%
“…On the other hand, when the table size becomes closer to be linear of n, data structures such as locality-sensitive hashing (LSH) [2,12] or data-dependent LSH [3,4] achieve a cell-probe complexity ofÕ(dn ρ ) with data structures of sizeÕ(n 1+ρ ) for some 0 < ρ < 1 depending on the metric and the approximation ratio. Compared to the Θ log log d log log log d bound of Chakrabarti and Regev, thẽ O(dn ρ ) cell-probe complexity is much worse.…”
mentioning
confidence: 99%
“…This makes all cell-probes in LSH parallelizable into one round of parallel memory accesses. And the more recent data-dependent LSH [3,4] surpasses the classic LSH in cell-probe complexity by being a little more adaptive: the algorithm retrieves a data-dependent hash function before making the second round of cell-probes, while the cell-probes in the second round are independent of each other. In contrast, the algorithm of Chakrabarti and Regev [10] is fully adaptive: Every cell-probe must wait for the information retrieved by the previous cell-probe to proceed.…”
mentioning
confidence: 99%