2018
DOI: 10.1002/cpe.4786
|View full text |Cite
|
Sign up to set email alerts
|

Energy‐based tuning of convolutional neural networks on multi‐GPUs

Abstract: Deep Learning (DL) applications are gaining momentum in the realm of Artificial Intelligence, particularly after GPUs have demonstrated remarkable skills for accelerating their challenging computational requirements. Within this context, Convolutional Neural Network (CNN) models constitute a representative example of success on a wide set of complex applications, particularly on datasets where the target can be represented through a hierarchy of local features of increasing semantic complexity. In most of the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 44 publications
0
5
0
Order By: Relevance
“…GPUs are extremely effective for several basic DL primitives, which include greatly parallel-computing operations such as activation functions, matrix multiplication, and convolutions [326][327][328][329][330]. Incorporating HBM-stacked memory into the up-to-date GPU models significantly enhances the bandwidth.…”
Section: Gpu-based Approachmentioning
confidence: 99%
“…GPUs are extremely effective for several basic DL primitives, which include greatly parallel-computing operations such as activation functions, matrix multiplication, and convolutions [326][327][328][329][330]. Incorporating HBM-stacked memory into the up-to-date GPU models significantly enhances the bandwidth.…”
Section: Gpu-based Approachmentioning
confidence: 99%
“…For comparison, a recent study [11] illustrated that performance per Watt on the Pascal GPU is 42 GFLOPs/Watt and on the Maxwell 23 GFLOPs/Watt for a similar machine learning problem. Whilst these results are significantly higher than those achieved in the micro-core LINPACK benchmark, crucially these two HPC grade GPUs draw a maximum of 250 Watts, whereas the power draw of the micro-cores used in our experiments was 0.90 Watts for the Epiphany and 0.18 Watts for the MicroBlaze.…”
Section: Experimentation Resultsmentioning
confidence: 99%
“…This machine combines a host dual core ARM A9 CPU, with 1 GB of RAM and the 16 core Epiphany-III. Due to limitations in the Parallella, whilst the theoretical off-chip bandwidth of the Epiphany III is 600 MB/s, the maximum obtainable in practice is 150 MB/s [11]. For MicroBlaze experiments we use the Pynq-II SBC, mounting a Xilinx Zynx-7020 and 512 MB RAM.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Li et al 57 analyze the energy and power behavior of CPUs and GPUs in deep learning training, and identifies the power‐hungry layers in the neural network by quantifying the energy consumption of each CNN layer. Castro et al 58 use two CNN models (i.e., ResNet, AlexNet) to perform combined energy and performance analysis on multi‐GPU settings. However, these two works mainly focus on the energy consumption of deep learning training.…”
Section: Related Workmentioning
confidence: 99%