2019
DOI: 10.1007/978-3-030-20518-8_26
|View full text |Cite
|
Sign up to set email alerts
|

Trainable Thresholds for Neural Network Quantization

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 6 publications
0
3
0
Order By: Relevance
“…Various quantization techniques have been proposed to make DNNs perform faster and fit larger networks on edge devices with limited storage capacity and energy budget [6][7][8]. An unfortunate consequence of quantization is the reduced accuracy, which can be tackled by increasing the network size, performing quantization only on parameters (and not on activations), or fine-tuning and re-training the network.…”
Section: Introductionmentioning
confidence: 99%
“…Various quantization techniques have been proposed to make DNNs perform faster and fit larger networks on edge devices with limited storage capacity and energy budget [6][7][8]. An unfortunate consequence of quantization is the reduced accuracy, which can be tackled by increasing the network size, performing quantization only on parameters (and not on activations), or fine-tuning and re-training the network.…”
Section: Introductionmentioning
confidence: 99%
“…[23] Recent research shows that post-training quantization that doesn't involve retraining or fine-tuning the full model can also effectively quantize the model. [24][25][26][27] Another critical issue of training convergence appears when utilizing the quantized version of neural networks during both training and testing. Chen et al proposed a hardware accelerator for convolutional neural networks and showed that the training precision should be selected as 32 bits instead of 16 bits to guarantee the convergence.…”
Section: General Approaches Towards Noise In Neural Networkmentioning
confidence: 99%
“…[ 23 ] Recent research shows that post‐training quantization that doesn't involve retraining or fine‐tuning the full model can also effectively quantize the model. [ 24–27 ]…”
Section: Introductionmentioning
confidence: 99%