2016
DOI: 10.1073/pnas.1513198113
|View full text |Cite
|
Sign up to set email alerts
|

Atoms of recognition in human and computer vision

Abstract: Discovering the visual features and representations used by the brain to recognize objects is a central problem in the study of vision. Recently, neural network models of visual object recognition, including biological and deep network models, have shown remarkable progress and have begun to rival human performance in some challenging tasks. These models are trained on image examples and learn to extract features and representations and to use them for categorization. It remains unclear, however, whether the r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

6
151
1

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 132 publications
(158 citation statements)
references
References 27 publications
6
151
1
Order By: Relevance
“…On the behavioural level, deep networks exhibit similar behaviour to humans (Hong, Yamins, Majaj, & DiCarlo, 2016;Kheradpisheh, Ghodrati, Ganjtabesh, & Masquelier, 2016b, 2016aKubilius, Bracci, & Op de Beeck, 2016;Majaj, Hong, Solomon, & DiCarlo, 2015) and are currently the best-performing model in explaining human eye-movements in free viewing paradigms (Kümmerer, Theis, & Bethge, 2014). Despite these advances, however, current DNNs still exhibit substantial differences in how they process and recognize visual stimuli (Linsley, Eberhardt, Sharma, Gupta, & Serre, 2017;Rajalingham et al, 2018;Ullman, Assif, Fetaya, & Harari, 2016), how they generalize to atypical category instances (Saleh, Elgammal, & Feldman, 2016), and how they perform under image manipulations, including reduced contrast and additive noise (Geirhos et al, 2017). Yet, the overall success clearly illustrates the power of DNN models for computational neuroscience.…”
Section: Brain-inspired Neural Network Models Are Revolutionising Artmentioning
confidence: 99%
“…On the behavioural level, deep networks exhibit similar behaviour to humans (Hong, Yamins, Majaj, & DiCarlo, 2016;Kheradpisheh, Ghodrati, Ganjtabesh, & Masquelier, 2016b, 2016aKubilius, Bracci, & Op de Beeck, 2016;Majaj, Hong, Solomon, & DiCarlo, 2015) and are currently the best-performing model in explaining human eye-movements in free viewing paradigms (Kümmerer, Theis, & Bethge, 2014). Despite these advances, however, current DNNs still exhibit substantial differences in how they process and recognize visual stimuli (Linsley, Eberhardt, Sharma, Gupta, & Serre, 2017;Rajalingham et al, 2018;Ullman, Assif, Fetaya, & Harari, 2016), how they generalize to atypical category instances (Saleh, Elgammal, & Feldman, 2016), and how they perform under image manipulations, including reduced contrast and additive noise (Geirhos et al, 2017). Yet, the overall success clearly illustrates the power of DNN models for computational neuroscience.…”
Section: Brain-inspired Neural Network Models Are Revolutionising Artmentioning
confidence: 99%
“…For instance, local contour information may be particularly useful for making subordinate-level category distinctions where the skeletons of objects are roughly the same 54,71 . Similarly, texture statistics and feature descriptions have been shown to be important indicators of both basic 11,72 and superordinate-level 73,74 object distinctions. And though it is currently unclear whether feedforward neural network models (such as the one tested here) incorporate global shape information 75,76 , more biologically plausible models with recurrent or generative architectures may begin to approximate human shape perception 68 .…”
Section: Discussionmentioning
confidence: 97%
“…An ensemble approach of Malisiewicz et al [27] predicts a class based on combined predictions from multiple linear Support Vector Machines (SVMs), each trained on only one example. Later development takes a deep learning approach [11][12] [13]. Although, most deep learning object detections do not require image features and can directly take image intensities as input.…”
Section: Customer Analysis Pipelinementioning
confidence: 99%