2018
DOI: 10.48550/arxiv.1806.05794
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

RAPIDNN: In-Memory Deep Neural Network Acceleration Framework

Mohsen Imani,
Mohammad Samragh,
Yeseong Kim
et al.

Abstract: Deep neural networks (DNN) have demonstrated effectiveness for various applications such as image processing, video segmentation, and speech recognition. Running state-ofthe-art DNNs on current systems mostly relies on either generalpurpose processors, ASIC designs, or FPGA accelerators, all of which suffer from data movements due to the limited on-chip memory and data transfer bandwidth. In this work, we propose a novel framework, called RAPIDNN, which performs neuron-tomemory transformation in order to accel… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 43 publications
0
5
0
Order By: Relevance
“…& Baseline / 16x CAP sharing architectures. 4-bit weight/activation quantization results in negligible decrease in functional performance (and actually better performance for ResNet) [51,52]. [18], the chip footprint of the 3D a-Cortex is ~16 / ~7 times smaller, while its energy efficiency is lower only a factor of ~5.4 / ~ 5.…”
Section: Comparison With Prior Workmentioning
confidence: 99%
See 1 more Smart Citation
“…& Baseline / 16x CAP sharing architectures. 4-bit weight/activation quantization results in negligible decrease in functional performance (and actually better performance for ResNet) [51,52]. [18], the chip footprint of the 3D a-Cortex is ~16 / ~7 times smaller, while its energy efficiency is lower only a factor of ~5.4 / ~ 5.…”
Section: Comparison With Prior Workmentioning
confidence: 99%
“…On the other hand, on the system level quite a few efforts were recently made to exploit the efficiency of MS operators to develop better DNN/RNN processor architectures [46][47][48][49][50][51][52]. For example, the ISAAC [46] and PUMA [47] architectures are 2D mesh structures of tiles where each tile contains several small (typically 128×128) ReRAM-based VMM units with their I/O peripheries.…”
Section: Comparison With Prior Workmentioning
confidence: 99%
“…Several ReRAM-based in-situ mixed-signal DNN accelerators such as ISAAC [18], Newton [19], PipeLayer [20], PRIME [17], PUMA [21], MultiScale [22], XNOR-RRAM [37], RapidDNN [38], have been proposed in recent years. These designs utilize a combination of analog and digital units to speed up the computation.…”
Section: A Reram Crossbar and Reram-based Dnn Accelerationmentioning
confidence: 99%
“…For example, heavy edge analytics in battery-driven self-driving cars, where safety and energy are critical considerations, can lead to their unexpected energy outage. For enhancing the energy efficiency of such devices, many energy-aware systolic array-based DNN accelerators have been recently developed [2] but their large size requirement for fast data processing poses meager energy gains in energy-constrained devices [3]. This problem can be addressed with approximate computing that trades the accuracy of an application-specific system, by exploiting its intrinsic error resilience, for energy savings [4].…”
Section: Introductionmentioning
confidence: 99%