Proceedings of the 2018 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays 2018
DOI: 10.1145/3174243.3174999
|View full text |Cite
|
Sign up to set email alerts
|

Exploration of Low Numeric Precision Deep Learning Inference Using Intel® FPGAs

Abstract: Convolutional neural networks (CNN) have been shown to maintain reasonable classification accuracy when quantized to lower precisions, however, quantizing to sub 8-bit activations and weights can result in classification accuracy falling below an acceptable threshold. Techniques exist for closing the accuracy gap of limited numeric precision networks typically by means of increasing computation. This results in a trade-off between throughput and accuracy and can be tailored for different networks through vario… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
20
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 21 publications
(20 citation statements)
references
References 17 publications
0
20
0
Order By: Relevance
“…For instance, the energy-delay-product of the posit EMAC with es = 0, on average, is 3× and 1.4× less than the energy-delay-product of the posit EMAC with es = 2 and es = 1, respectively. On the other hand, Fixed Float Posit the average performance of DNN inference with es = 1 for the posit EMAC among the five datasets and [5,7] bit-precision is 2% and 4% percent better than with es = 2 and es = 0, respectively. Thus, Deep Positron equipped with the posit (es = 1) EMAC has a better trade-off between energy-delay-product and accuracy for [5,7] bits.…”
Section: Exploiting the Posit Es Parametermentioning
confidence: 92%
See 1 more Smart Citation
“…For instance, the energy-delay-product of the posit EMAC with es = 0, on average, is 3× and 1.4× less than the energy-delay-product of the posit EMAC with es = 2 and es = 1, respectively. On the other hand, Fixed Float Posit the average performance of DNN inference with es = 1 for the posit EMAC among the five datasets and [5,7] bit-precision is 2% and 4% percent better than with es = 2 and es = 0, respectively. Thus, Deep Positron equipped with the posit (es = 1) EMAC has a better trade-off between energy-delay-product and accuracy for [5,7] bits.…”
Section: Exploiting the Posit Es Parametermentioning
confidence: 92%
“…On the other hand, Fixed Float Posit the average performance of DNN inference with es = 1 for the posit EMAC among the five datasets and [5,7] bit-precision is 2% and 4% percent better than with es = 2 and es = 0, respectively. Thus, Deep Positron equipped with the posit (es = 1) EMAC has a better trade-off between energy-delay-product and accuracy for [5,7] bits. For 8-bit, the results suggest that es = 1 is a better fit for energy-efficient applications and es = 2 for accuracy-dependent applications.…”
Section: Exploiting the Posit Es Parametermentioning
confidence: 92%
“…The training of such models requires high computational power, however, once trained the network can work in realtime with even 15fps rate of processing power. The work in literature [61][62][63] mentions details about the training and computation power.…”
Section: Algorithmsmentioning
confidence: 99%
“…At present, there are two solutions to accelerate the use of CNNs. One is to reduce the computational complexity of the neural network; many such methods have been proposed that maintain the accuracy, including quantification, tailoring, sparsity and fast convolution [7][8][9][10][11][12][13]. The other solution is to use a high-performance, low-power hardware accelerator.…”
Section: Introductionmentioning
confidence: 99%