2014 IEEE 55th Annual Symposium on Foundations of Computer Science 2014
DOI: 10.1109/focs.2014.68
|View full text |Cite
|
Sign up to set email alerts
|

Spectral Approaches to Nearest Neighbor Search

Abstract: We study spectral algorithms for the high-dimensional Nearest Neighbor Search problem (NNS). In particular, we consider a semi-random setting where a dataset P in R d is chosen arbitrarily from an unknown subspace of low dimension k d, and then perturbed by fully d-dimensional Gaussian noise. We design spectral NNS algorithms whose query time depends polynomially on d and log n (where n = |P |) for large ranges of k, d and n. Our algorithms use a repeated computation of the top PCA vector/subspace, and are eff… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
4
3

Relationship

2
5

Authors

Journals

citations
Cited by 15 publications
(15 citation statements)
references
References 47 publications
0
15
0
Order By: Relevance
“…We stress that this approach gives improvement for worstcase datasets, which is somewhat unexpected. To put this into a perspective: if one were to assume that the dataset has some special structure, it would be more natural to expect speed-ups with data-dependent hashing: such hashing may adapt to the special structure, perhaps implicitly, as was done in, say, [9,29,1]. However, in our setting there is no assumed structure to adapt to, and hence it is unclear why data-dependent hashing shall help.…”
Section: Introductionmentioning
confidence: 92%
See 3 more Smart Citations
“…We stress that this approach gives improvement for worstcase datasets, which is somewhat unexpected. To put this into a perspective: if one were to assume that the dataset has some special structure, it would be more natural to expect speed-ups with data-dependent hashing: such hashing may adapt to the special structure, perhaps implicitly, as was done in, say, [9,29,1]. However, in our setting there is no assumed structure to adapt to, and hence it is unclear why data-dependent hashing shall help.…”
Section: Introductionmentioning
confidence: 92%
“…First, we show how to achieve success probability n −ρ , query time n oc(1) , and space and preprocessing time n 1+oc(1) , where ρ = 1 2c 2 −1 + oc (1). Finally, to obtain the final result, one then builds O n ρ copies of the above data structure to amplify the probability of success to 0.99 (as explained in Remark 2).…”
Section: The Data Structurementioning
confidence: 97%
See 2 more Smart Citations
“…Most such examples include the algorithms that assume some additional structure in the dataset: such as some notion of low intrinsic dimension [KR02, CNBM01, KL04, BKL06, IN07, Cla06, DF08], or low dimensional data-set with high-dimensional noise [AAKK14]. Most relevant to us is the work of [DS13], which, while mostly focusing on the low intrinsic dimensional datasets, give a generic bound for the worst-case datasets as well.…”
Section: Related Workmentioning
confidence: 99%