2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Volume 2 (CVPR'06)
DOI: 10.1109/cvpr.2006.264
|View full text |Cite
|
Sign up to set email alerts
|

Scalable Recognition with a Vocabulary Tree

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

11
2,726
0
12

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 2,890 publications
(2,749 citation statements)
references
References 16 publications
11
2,726
0
12
Order By: Relevance
“…We measure the retrieval error rate as the percentage of query images not correctly retrieved and verified with our pipeline. We briefly describe our retrieval pipeline, which is similar to other the state-of-the-art systems, such as [23,24,25,26].…”
Section: Database Retrievalmentioning
confidence: 99%
See 2 more Smart Citations
“…We measure the retrieval error rate as the percentage of query images not correctly retrieved and verified with our pipeline. We briefly describe our retrieval pipeline, which is similar to other the state-of-the-art systems, such as [23,24,25,26].…”
Section: Database Retrievalmentioning
confidence: 99%
“…Using these descriptors, we train a 10 6 leaf, 6 level, vocabulary tree [23]. We use symmetric KL-divergence as the distance measure for both training and querying, since it performs better than L 2 -norm for HoG descriptors [27].…”
Section: Database Retrievalmentioning
confidence: 99%
See 1 more Smart Citation
“…To build a vocabulary of visual words, interest regions in the images are detected with Hessian-affine detector [8], which provides good performance [9] and is widely used in visual word-based studies because of its insensitiveness to affine transformations such as scaling, reflection, rotation, etc. These regions are described in 128-dimension SIFT descriptors and then clustered by a hierarchical k-means algorithm [10], each cluster representing a visual word. Then each image is represented in a bag of visual words.…”
Section: B Detecting Reused Visual Elementsmentioning
confidence: 99%
“…These features are robust to arbitrary changes in viewpoints. Then, hierarchical k-means clustering [19] is applied to the features, to group them based on their similarity. The result of the hierarchical clustering is used for the fast approximation of the nearest neighbor search, to efficiently resolve feature matching.…”
Section: Object Duplicate Detectionmentioning
confidence: 99%