2019
DOI: 10.1109/tc.2018.2879434
|View full text |Cite
|
Sign up to set email alerts
|

SparCE: Sparsity Aware General-Purpose Core Extensions to Accelerate Deep Neural Networks

Abstract: Deep Neural Networks (DNNs) have emerged as the method of choice for solving a wide range of machine learning tasks. Satiating the enormous growth in computational demand posed by DNNs is a key challenge for computing system designers and has most commonly been addressed through the design of custom accelerators. However, these specialized accelerators that utilize large quantities of multiply-accumulate units and on-chip memory are prohibitive in many design scenarios (e.g., wearable devices and IoT sensors),… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
25
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 29 publications
(25 citation statements)
references
References 32 publications
0
25
0
Order By: Relevance
“…The AFTx06 implements an RV32IM instruction set, integer multiplication and division, an integrated core local interrupt (CLINT) controller and a platform-level interrupt controller (PLIC), and it is expected to operate with a 100-MHz clock. A key feature of this microprocessor is the sparsity-aware core extensions (SparCE) architecture [2], a component that exploits sparsity in convolution arithmetic, allowing extraneous instructions (such as multiplication by zero; common in rectified linear unit functions) to be skipped at run time. The architecture has been designed to improve both the speed and power consumption of this common machine learning calculation.…”
Section: Sparsity-optimized Risc-v Soc: Combining Undergraduate Education and Soc Researchmentioning
confidence: 99%
“…The AFTx06 implements an RV32IM instruction set, integer multiplication and division, an integrated core local interrupt (CLINT) controller and a platform-level interrupt controller (PLIC), and it is expected to operate with a 100-MHz clock. A key feature of this microprocessor is the sparsity-aware core extensions (SparCE) architecture [2], a component that exploits sparsity in convolution arithmetic, allowing extraneous instructions (such as multiplication by zero; common in rectified linear unit functions) to be skipped at run time. The architecture has been designed to improve both the speed and power consumption of this common machine learning calculation.…”
Section: Sparsity-optimized Risc-v Soc: Combining Undergraduate Education and Soc Researchmentioning
confidence: 99%
“…SparCE [40] skips ineffectual code blocks based on a sparse input. It annotates skippable code blocks in software and tests conditions in hardware.…”
Section: Related Workmentioning
confidence: 99%
“…Indeed, prior efforts spanning hardware to software and algorithms have exploited sparsity to eliminate computation or data transfers at different points in DNN computations. Most of these efforts, though, require hardware changes [3,7,13,34,38,40,57] and/or apply only to inference [3,7,13,15,34,35,50,53,57]. This is not ideal, since most of real-world DNN computations are performed on conventional CPUs and GPUs [4,16,33,51], and significant time goes into training.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Finally, we compare our work on exploiting zeros in modern gaming applications with prior art on leveraging sparsity (i.e., memory loads and computations returning zeros) in deep neural networks (DNNs). Several researchers have exploited sparsity to improve DNN performance in accelerator architectures [1,34,50,65], GPUs [51], and general-purpose processors [56] through hardware enhancements. Like these efforts, Zeroploit seeks to improve performance by leveraging zero valued operands.…”
Section: Related Workmentioning
confidence: 99%