2015
DOI: 10.1145/2872887.2750389
|View full text |Cite
|
Sign up to set email alerts
|

ShiDianNao

Abstract: In recent years, neural network accelerators have been shown to achieve both high energy efficiency and high performance for a broad application scope within the important category of recognition and mining applications.Still, both the energy efficiency and performance of such accelerators remain limited by memory accesses. In this paper, we focus on image applications, arguably the most important category among recognition and mining applications. The neural networks which are state-of-the-art for these appli… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
19
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 148 publications
(19 citation statements)
references
References 47 publications
0
19
0
Order By: Relevance
“…Krizhevsky, A. et al used a local response normalization (LRN) operation in AlexNet to reduce the error rates of top-1 and top-5 by 1.4% and 1.2% [24]. Du, Z. et al added local response normalization (LRN) and local contrast normalization (LCN) in the design of ShiDianNao, which improved the recognition accuracy, but increased the computational and hardware complexity [40]. The batch normalization proposed by Ioffe, S. et al in 2015 is widely used in the deep neural network, which effectively accelerates the speed of training and convergence [51].…”
Section: Normalization Layermentioning
confidence: 99%
See 1 more Smart Citation
“…Krizhevsky, A. et al used a local response normalization (LRN) operation in AlexNet to reduce the error rates of top-1 and top-5 by 1.4% and 1.2% [24]. Du, Z. et al added local response normalization (LRN) and local contrast normalization (LCN) in the design of ShiDianNao, which improved the recognition accuracy, but increased the computational and hardware complexity [40]. The batch normalization proposed by Ioffe, S. et al in 2015 is widely used in the deep neural network, which effectively accelerates the speed of training and convergence [51].…”
Section: Normalization Layermentioning
confidence: 99%
“…proposed ShiDianNao based on 2-D mesh topology structure for image recognition applications near to sensors, and reduced memory usage through weight sharing [40]. Zhang, C. et al designed a CNN accelerator based on the adder tree structure by quantitative analysis of memory bandwidth required for throughput [41].…”
Section: Introductionmentioning
confidence: 99%
“…DaDianNao supports convolution, pooling, class er and LRN layers, and when using a 64-node architecture, it achieves more than 2000x accelerations in convolution computation compared to GPU baselines. ShiDianNao [8] focuses on accelerating convolution operations in embedded applications, and supports pooling, classi cation, and normalization layers as well. ShiDianNao uses inter-PE data propagation to reduce memory access in convolution, which makes it high energy e ciency.…”
Section: Introductionmentioning
confidence: 99%
“…Several libraries and frameworks have been developed for the implementation of DNNs via GPUs; these include Theano (which is a Python library) [ 20 ] and Caffe (a deep learning framework) [ 21 ], Tensor Flow and Chianer (Python-based deep learning frameworks) [ 22 , 23 ]. Some DNNs, such as some MLPs [ 24 , 25 ], RBMs [ 26 , 27 ], and CNNs [ 28 – 31 ], have been developed as dedicated chips. One of the report uses RBMs for training and AEs for inference [ 32 ].…”
Section: Introductionmentioning
confidence: 99%