Proceedings of the Second Annual ACM Conference on Multimedia Systems 2011
DOI: 10.1145/1943552.1943568
|View full text |Cite
|
Sign up to set email alerts
|

The stanford mobile visual search data set

Abstract: We survey popular data sets used in computer vision literature and point out their limitations for mobile visual search applications. To overcome many of the limitations, we propose the Stanford Mobile Visual Search data set. The data set contains camera-phone images of products, CDs, books, outdoor landmarks, business cards, text documents, museum paintings and video clips. The data set has several key characteristics lacking in existing data sets: rigid objects, widely varying lighting conditions, perspectiv… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
72
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 129 publications
(73 citation statements)
references
References 17 publications
1
72
0
Order By: Relevance
“…In order to evaluate the performance of the proposed matching method, the Stanford Mobile Visual Search (SMVS) dataset from Stanford University was used [21]. SMVS includes images of CDs, DVDs, books, paintings, and video clips, and is currently the standard image set for performance evaluation of image matching under MPEG-7 CDVS.…”
Section: Experiments Resultsmentioning
confidence: 99%
“…In order to evaluate the performance of the proposed matching method, the Stanford Mobile Visual Search (SMVS) dataset from Stanford University was used [21]. SMVS includes images of CDs, DVDs, books, paintings, and video clips, and is currently the standard image set for performance evaluation of image matching under MPEG-7 CDVS.…”
Section: Experiments Resultsmentioning
confidence: 99%
“…Without pretending to be exhaustive in our overview, we mention that ALOI dataset containing 1000 objects [1], the ImageNet dataset containing 200 categories of objects [2], the dataset from the PASCAL 2012 challenge containing 20 classes of objects [3] and probably the closest to our application, the Stanford Mobile visual search data set contains 23 different objects such as books, CD covers, DVD covers and common objects [4].…”
Section: A State-of-the-art In the Package Databasesmentioning
confidence: 99%
“…Their approach computes Fisher vectors by aggregating ORB descriptors under the assumption that they are generated from a Bernoulli mixture model [12]. Fisher vectors composed of ORB descriptors were compared with bag of binary words [13] on the Stanford mobile visual search dataset (SMVSD) [14]. Although they were found to be better than the competitor, they did not compare their approach with the typical VLAD.…”
Section: Local Descriptor Extraction and Descriptor Aggregationmentioning
confidence: 99%
“…Next, we studied how much our framework accelerates the image retrieval process and reduces the amount of storage required compared with other approaches. We conducted experiments on three datasets, holidays [24], ukbench [25], SMVSD † [14]. All images were resized to less than 480 pixels beforehand.…”
Section: Overview and Protocolmentioning
confidence: 99%