Proceedings of the 37th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems 2018
DOI: 10.1145/3196959.3196976
|View full text |Cite
|
Sign up to set email alerts
|

Distance-Sensitive Hashing

Abstract: Locality-sensitive hashing (LSH) is an important tool for managing high-dimensional noisy or uncertain data, for example in connection with data cleaning (similarity join) and noise-robust search (similarity search). However, for a number of problems the LSH framework is not known to yield good solutions, and instead ad hoc solutions have been designed for particular similarity and distance measures. For example, this is true for output-sensitive similarity search/join, and for indexes supporting annulus queri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 16 publications
(12 citation statements)
references
References 52 publications
0
12
0
Order By: Relevance
“…Approximating Log-convex Functions via Distance Sensitive Hashing. We show that this is indeed possible for log-convex functions of the inner product by utilizing a family of hashing schemes introduced recently by Aumuller et al [15], referred to as Distance Sensitive Hashing (DSH). This family is defined through two parameters γ 0 and s > 0, with collision probability p γ,s (ρ) having the following dependence on the inner product ρ x, y between two vectors x, y ∈ S d−1…”
Section: Do There Exist Design Principles For {W T } and {P T } That mentioning
confidence: 87%
See 1 more Smart Citation
“…Approximating Log-convex Functions via Distance Sensitive Hashing. We show that this is indeed possible for log-convex functions of the inner product by utilizing a family of hashing schemes introduced recently by Aumuller et al [15], referred to as Distance Sensitive Hashing (DSH). This family is defined through two parameters γ 0 and s > 0, with collision probability p γ,s (ρ) having the following dependence on the inner product ρ x, y between two vectors x, y ∈ S d−1…”
Section: Do There Exist Design Principles For {W T } and {P T } That mentioning
confidence: 87%
“…To analyze the collision probability of the DSH scheme we closely follow the proof of Aumuller et al [15] with the difference that we use Proposition 9.2 to bound bi-variate Gaussian integrals.…”
Section: Distance Sensitive Hashing On the Unit Spherementioning
confidence: 99%
“…An interesting open question is to investigate the applicability of our data structures for problems like discrimination discovery [56], diversity in recommender systems [3], privacy preserving similarity search [65], and estimation of kernel density [22]. Moreover, it would be interesting to investigate techniques for providing incentives (i.e., reverse discrimination [56]) to prevent discrimination: an idea could be to merge the data structures in this paper with distance-sensitive hashing functions in [13], which allow to implement hashing schemes where the collision probability is an (almost) arbitrary function of the distance. Further, the techniques presented here require a manual trade-of between the performance of the LSH part and the additional running time contribution from inding the near points among the non-far points.…”
Section: Discussionmentioning
confidence: 99%
“…More recently, researchers have begun leveraging LSH techniques to solve problems beyond ANN, extending their domain to applications around density estimation for high-dimensional models. For example [ACPS17] generalizes nearest neighbor LSH hash functions to be sensitive to custom distance ranges. [AAP17] builds many different parameterized versions of the prototypical LSH hash tables and adaptively probes them for spherical range reporting.…”
Section: Related Literaturementioning
confidence: 99%