2010 18th Euromicro Conference on Parallel, Distributed and Network-Based Processing 2010
DOI: 10.1109/pdp.2010.43
|View full text |Cite
|
Sign up to set email alerts
|

Performance and Scalability of GPU-Based Convolutional Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
86
0
1

Year Published

2011
2011
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 202 publications
(95 citation statements)
references
References 12 publications
1
86
0
1
Order By: Relevance
“…Most previous GPUbased CNN implementations [11], [12] were hard-coded to satisfy GPU hardware constraints, whereas ours is flexible and fully online (i.e., weight updates after each image). Other flexible implementations [13] are not fully exploiting the latest GPUs. It allows for training large CNNs within days instead of months, such that we can investigate the influence of various structural parameters by exploring large parameter spaces [14] and performing error analysis on repeated experiments.…”
Section: Introductionmentioning
confidence: 99%
“…Most previous GPUbased CNN implementations [11], [12] were hard-coded to satisfy GPU hardware constraints, whereas ours is flexible and fully online (i.e., weight updates after each image). Other flexible implementations [13] are not fully exploiting the latest GPUs. It allows for training large CNNs within days instead of months, such that we can investigate the influence of various structural parameters by exploring large parameter spaces [14] and performing error analysis on repeated experiments.…”
Section: Introductionmentioning
confidence: 99%
“…Moreover, a nonlinear activation (e.g. sigmoid, hyperbolic tangent, rectified linear units) function is taken outside the convolutional layer to strengthen the non-linearity [44]. Specifically, the major operations performed in the CNN can be summarized as:…”
Section: A Convolutional Neural Network (Cnn)mentioning
confidence: 99%
“…The CNN also provides partial resistance and robustness to geometric distortions and transformations, and other 2D shape variations [2]. Hence, the CNN is specifically designed to cope with shortcomings of the traditional feature extractor that is characterized by being static, is designed independently of the trainable classifier, and is not part of training procedure [3]. A final benefit of CNNs is that they are relatively easier to train since they have fewer parameters than fully connected MLP neural networks with the same number of hidden layers.…”
Section: Raw Inputmentioning
confidence: 99%
“…The concepts of local receptive field, weight sharing, and spatial subsampling mentioned above are the three principle architectural ideas behind the design of a CNN [2,6]. In weight sharing topology, all neurons in a feature map use the same incoming set of weights (kernel weights), and feature extraction is performed by convolving the image with these kernels [3,11].…”
Section: Background On Convolutional Neural Networkmentioning
confidence: 99%