1998
DOI: 10.1145/293347.293348
|View full text |Cite
|
Sign up to set email alerts
|

An optimal algorithm for approximate nearest neighbor searching fixed dimensions

Abstract: , or permissions@acm.org. · S. Arya, et al.Consider a set S of n data points in real d-dimensional space, R d , where distances are measured using any Minkowski metric. In nearest neighbor searching we preprocess S into a data structure, so that given any query point q ∈ R d , the closest point of S to q can be reported quickly. Given any positive real , a data point p is a (1 + )-approximate nearest neighbor of q if its distance from q is within a factor of (1 + ) of the distance to the true nearest neighbor.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

6
1,724
1
7

Year Published

2001
2001
2017
2017

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 2,094 publications
(1,738 citation statements)
references
References 58 publications
6
1,724
1
7
Order By: Relevance
“…The algorithm incrementally selects pivotp that maximizes the sum of the lower bounds on distances between data objects [16] as shown in Eqs. (3) and (4). The distance lower bound is derived by the triangle inequality applied to the triangle with the pivot and two objects.…”
Section: Pivot Generation and Data Partitioning Algorithmsmentioning
confidence: 99%
See 1 more Smart Citation
“…The algorithm incrementally selects pivotp that maximizes the sum of the lower bounds on distances between data objects [16] as shown in Eqs. (3) and (4). The distance lower bound is derived by the triangle inequality applied to the triangle with the pivot and two objects.…”
Section: Pivot Generation and Data Partitioning Algorithmsmentioning
confidence: 99%
“…From a perspective of resultant search accuracy, similarity search methods are classified into three main categories of exact, approximate, and heuristic search. Lately, to solve similarity search problems for large-scale data sets, approximate search methods that guarantee some resultant accuracy have been studied with considerable effort, which contain those based on a tree-type index [4] and locality-sensitive hashing (LSH) family [5]- [7]. In contrast, exact methods have received interest for a long time in the application domains where a data set has relatively low intrinsic dimensionality [3].…”
Section: Introductionmentioning
confidence: 99%
“…Each of the feature vectors extracted are then classified by the k-NN algorithm, which produces a hypothesis label for each pixel, using the fast approximate nearest neighbor search, based on kdtrees [34]. The result of the classification stage is a binary map representing tissue types.…”
Section: Fully-automated Classificationmentioning
confidence: 99%
“…First, nearest neighbor search algorithms that are not computationally exhaustive degrade as a function of the dimension of the data. For example, the popular ApproximateNearest-Neighbor algorithm [1] computes a (1 + ε)-approximate nearest neighbor of a point in O((H 1+6H/ε) H log N) time. This approach is far too expensive for high-dimensional data where H = 28, 374 as it is in one of our test datasets in section V. Second, measurement noise can destabilize the topology of such graphs making the results sensitive to the parameters of the algorithm used to construct the graph.…”
Section: Graph Layout Algorithmsmentioning
confidence: 99%