2020
DOI: 10.1038/s41598-020-59175-0
|View full text |Cite
|
Sign up to set email alerts
|

Orthogonal Representations of Object Shape and Category in Deep Convolutional Neural Networks and Human Visual Cortex

Abstract: Deep Convolutional Neural Networks (CNNs) are gaining traction as the benchmark model of visual object recognition, with performance now surpassing humans. While CNNs can accurately assign one image to potentially thousands of categories, network performance could be the result of layers that are tuned to represent the visual shape of objects, rather than object category, since both are often confounded in natural images. Using two stimulus sets that explicitly dissociate shape from category, we correlate thes… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
43
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 72 publications
(52 citation statements)
references
References 39 publications
3
43
0
Order By: Relevance
“…For the stimulus set of Bracci & Op de Beeck (2016), all three training regimes of MobileNet preferred shape over category. In addition, they also showed a smaller, yet significant correlation with the category-related behavioural similarities, as it was also shown before for other networks trained with unfiltered images by Zeman et al (2020). There are no meaningful effects of the training regime, so the extent and timing of training with low-pass filtered images does not affect the presence of a multi-feature representation in a CNN.…”
Section: Comparison With Human Performance For Low-pass Filtered Imagessupporting
confidence: 68%
See 3 more Smart Citations
“…For the stimulus set of Bracci & Op de Beeck (2016), all three training regimes of MobileNet preferred shape over category. In addition, they also showed a smaller, yet significant correlation with the category-related behavioural similarities, as it was also shown before for other networks trained with unfiltered images by Zeman et al (2020). There are no meaningful effects of the training regime, so the extent and timing of training with low-pass filtered images does not affect the presence of a multi-feature representation in a CNN.…”
Section: Comparison With Human Performance For Low-pass Filtered Imagessupporting
confidence: 68%
“…Bracci et al (2019) found that CNNs classify objects by animacy rather than appearance, and this bias towards animacy is also demonstrated in human judgements and brain representations. Zeman et al (2020) found that CNNs represent category independently from shape in object images, similarly to human object recognition areas. All together, these studies and more, have demonstrated the strength and breadth of applications in using CNNs as a model of human vision.…”
Section: Introductionmentioning
confidence: 89%
See 2 more Smart Citations
“…Furthermore, there is increasing evidence that, even though most artificial neural networks largely lack biological plausibility, they are nevertheless well-suited for modeling brain function. A number of recent studies have shown striking similarities in the processing and representational dynamics between artificial neural networks and the brain (Cichy et al, 2016;Zeman et al, 2020). For instance, in deep neural networks trained on visual object recognition, the spontaneous emergence of number detectors (Nasr et al, 2019), solid shape coding (Srinath et al, 2020), or center-periphery spatial organization (Mohsenzadeh et al, 2020) was observed.…”
Section: The Global Workpace Theorymentioning
confidence: 99%