Big Data 2016
DOI: 10.1016/b978-0-12-805394-2.00004-0
|View full text |Cite
|
Sign up to set email alerts
|

Deep Learning and Its Parallelization

et al.
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
15
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 23 publications
(16 citation statements)
references
References 5 publications
0
15
0
Order By: Relevance
“…A deep CNN consists of an input layer that contains image data of m training examples, multiple hidden layers that compute features from input images and an output layer, which classifies the learned images. Deep learning models employ non-linear transformation functions to solve complex large-scale problems (Reyes et al, 2015;Li et al, 2016;Badejo et al, 2018). As shown in Figure 1, the hidden layers consist of stacked convolution layers that convolve using a Rectified Linear Unit (ReLU) activation (or transfer) function, as well as a pooling layer, which reduces the dimension of the convoluted image.…”
Section: Convolutional Neural Networkmentioning
confidence: 99%
See 1 more Smart Citation
“…A deep CNN consists of an input layer that contains image data of m training examples, multiple hidden layers that compute features from input images and an output layer, which classifies the learned images. Deep learning models employ non-linear transformation functions to solve complex large-scale problems (Reyes et al, 2015;Li et al, 2016;Badejo et al, 2018). As shown in Figure 1, the hidden layers consist of stacked convolution layers that convolve using a Rectified Linear Unit (ReLU) activation (or transfer) function, as well as a pooling layer, which reduces the dimension of the convoluted image.…”
Section: Convolutional Neural Networkmentioning
confidence: 99%
“…The Fully Connected (FC) layer connects each input to the next layer (classification layer) from the previous layer. Significant achievements have been made through the application of deep learning models in image processing, natural language processing and speech recognition tasks, thereby paving the way for more predictive analysis of big data (Li et al, 2016). The description of the aforementioned layers can be mathematically represented as shown in Equation ( 1)-( 7):…”
Section: Convolutional Neural Networkmentioning
confidence: 99%
“…There are three prominent strategies to partition the learning phase of a model: partitioning by input samples (data parallelism), by network structure (model parallelism), and by layer (pipelining). Data parallelism can be easily implemented, and it is, therefore, the most widely used implementation strategy on multi-GPUs (Li et al (2016)). We have explored this option focusing on CPUs only, where each core utilises the same sparse model to train on di erent data subsets.…”
Section: Parallel Training Of Deep Neural Networkmentioning
confidence: 99%
“…The key consideration to take advantage of their parallelization is to know how to divide the GPUs' tasks. Hence, three approaches should be considered if we want to train parallelized models: data parallelism, model parallelism, and data-model parallelism [79].…”
Section: Parallelization Of Neural Networkmentioning
confidence: 99%