2012 IEEE Conference on Computer Vision and Pattern Recognition 2012
DOI: 10.1109/cvpr.2012.6248110
|View full text |Cite
|
Sign up to set email alerts
|

Multi-column deep neural networks for image classification

Abstract: Traditional methods of computer vision and machine learning cannot match human performance on tasks such as the recognition of handwritten digits or traffic signs. Our biologically plausible, wide and deep artificial neural network architectures can. Small (often minimal) receptive fields of convolutional winner-take-all neurons yield large network depth, resulting in roughly as many sparsely connected neural layers as found in mammals between retina and visual cortex. Only winner neurons are trained. Several … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

11
1,790
2
18

Year Published

2015
2015
2022
2022

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 2,889 publications
(1,821 citation statements)
references
References 25 publications
11
1,790
2
18
Order By: Relevance
“…An ensemble of GPU-MPCNNs was also the first method to achieve human-competitive performance (around 0.2%) on MNIST (Ciresan et al, 2012c). This represented a dramatic improvement, since by then the MNIST record had hovered around 0.4% for almost a decade (Sec.…”
Section: : Mpcnns On Gpu Achieve Superhuman Vision Performancementioning
confidence: 99%
“…An ensemble of GPU-MPCNNs was also the first method to achieve human-competitive performance (around 0.2%) on MNIST (Ciresan et al, 2012c). This represented a dramatic improvement, since by then the MNIST record had hovered around 0.4% for almost a decade (Sec.…”
Section: : Mpcnns On Gpu Achieve Superhuman Vision Performancementioning
confidence: 99%
“…For example, it is unclear what in the images the learned networks actually look at and how the input image is represented in them. This is in stark contrast with the recent accelerated improvements of methods for training deep networks [2,8,13,6]. This lack of understanding leads to real problems; for example, a lot of trial-and-errors are necessary when designing the network architecture for each problem.…”
Section: Introductionmentioning
confidence: 49%
“…Correct identification of a digit '0' was highest at p = 0.985 and correct identification of a '5' lowest at p = 0.874. The five most significant error confusions are (2,8), (9,4), (4,9), (5,8), (5,3).…”
Section: Implementation and Resultsmentioning
confidence: 99%
“…2 The Convnet chosen uses a hierarchy of three macro-levels, each level comprises a convolutional layer, rectified linear unit layer, max pool layer, and drop out layer. At the top of all this, there is an output processing layer termed 'softmax' or normalised exponential, making 13 layers in total.…”
Section: Performance Using a Deep Network And Independent Data Sourcesmentioning
confidence: 99%