2016 IEEE International Solid-State Circuits Conference (ISSCC) 2016
DOI: 10.1109/isscc.2016.7418008
|View full text |Cite
|
Sign up to set email alerts
|

14.6 A 1.42TOPS/W deep convolutional neural network recognition processor for intelligent IoE systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
49
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 119 publications
(50 citation statements)
references
References 6 publications
1
49
0
Order By: Relevance
“…One big differentiating point between these platforms are their assumptions in terms of algorithmic and arithmetic accuracy. Jaehyeong et al [29] rely on 24bit fixed-point arithmetics, but they approximate weights using a low-dimensional representation based on PCA. Most other works use either 16 bits [30], [31] or 12 bits [26].…”
Section: B Low-power Cnn Hardware Ipsmentioning
confidence: 99%
See 1 more Smart Citation
“…One big differentiating point between these platforms are their assumptions in terms of algorithmic and arithmetic accuracy. Jaehyeong et al [29] rely on 24bit fixed-point arithmetics, but they approximate weights using a low-dimensional representation based on PCA. Most other works use either 16 bits [30], [31] or 12 bits [26].…”
Section: B Low-power Cnn Hardware Ipsmentioning
confidence: 99%
“…e Weights produced on-chip from a small set of PCA bases to save area/power. No evaluation on the general validity of this approach is presented in [29]. f Performance & power of inference engines only, estimating they are responsible for 20% of total power.…”
Section: A System-on-chip Operating Modesmentioning
confidence: 99%
“…KU Leuven's accelerator [16] is a SIMD array system with dynamic voltage and bit precision control, aiming for low-power mobile applications. The accelerator proposed by KAIST [17] is a CNN accelerator which employs principal component analysis for the weights of convolutional layers to minimize the data size read from external memory. ShiDianNao [18] and its previous work DaDianNao [19] also retain the weight values in the internal buffers and employ spatial-mapped neural function unit.…”
Section: Related Workmentioning
confidence: 99%
“…In academia, three representative works at the architectural level are Eyeriss [23], EIE [24], and the DianNao family [25][26][27], which focus specifically on the convolutional layers, the fully-connected layers, and the memory design/organization, respectively. There are a number of recent tapeouts of hardware deep learning systems [23,[28][29][30][31][32][33].…”
Section: Introductionmentioning
confidence: 99%