2018
DOI: 10.1109/tpami.2017.2723009
|View full text |Cite
|
Sign up to set email alerts
|

Places: A 10 Million Image Database for Scene Recognition

Abstract: The rise of multi-million-item dataset initiatives has enabled data-hungry machine learning algorithms to reach near-human semantic classification performance at tasks such as visual object and scene recognition. Here we describe the Places Database, a repository of 10 million scene photographs, labeled with scene semantic categories, comprising a large and diverse list of the types of environments encountered in the world. Using the state-of-the-art Convolutional Neural Networks (CNNs), we provide scene class… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
2,311
3
12

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 3,099 publications
(2,329 citation statements)
references
References 33 publications
3
2,311
3
12
Order By: Relevance
“… FABMAP [1]: This is a state-of-the-art place recognition algorithm built on top of handcrafted features  SeqSLAM [14]: This is a sequence-based place recognition approach, which have demonstrated state-of-the-art performances on mapping environments across seasons, weather conditions and different times of a day.  Places365 [22]: This is a CNN-based scene recognition model trained to recognize 365 scene types. The model was trained on over two million pictures.…”
Section: E Comparison Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“… FABMAP [1]: This is a state-of-the-art place recognition algorithm built on top of handcrafted features  SeqSLAM [14]: This is a sequence-based place recognition approach, which have demonstrated state-of-the-art performances on mapping environments across seasons, weather conditions and different times of a day.  Places365 [22]: This is a CNN-based scene recognition model trained to recognize 365 scene types. The model was trained on over two million pictures.…”
Section: E Comparison Methodsmentioning
confidence: 99%
“…Rather than recognizing specific places, networks have also been trained for recognizing the types of places [22]. However, this scene recognition task is different in nature from visual place recognition: image under the same scene category can come from different places.…”
Section: A Visual Place Recognition With Cnnsmentioning
confidence: 99%
“…The Places dataset [44] consists Urban Sci. 2018, 2, 78 7 of 19 of around 10 million images depicting indoor and outdoor views.…”
Section: Scene Classificationmentioning
confidence: 99%
“…As the foundation for our spatial attention model, we used a goal-directed CNN with a deep architecture 10 trained for scene categorization 18,19 . This architecture consists of 18 spatiallyselective layers that compute alternating convolution and non-linear max-pooling operations (Fig.…”
Section: Spatial Attention Model Definitionmentioning
confidence: 99%
“…To facilitate straightforward interpretation of which factors drive significant prediction of eye movements from spatial priority maps reconstructed from fMRI activity, these steps were excluded from our model. 10,18,19 . (b) Unit activity was extracted from the five pooling layers to sample activity from across the CNN hierarchy.…”
Section: Spatial Attention Model Definitionmentioning
confidence: 99%