2024
DOI: 10.2478/jee-2024-0051
|View full text |Cite
|
Sign up to set email alerts
|

Accuracy degradation aware bit rate allocation for layer-wise uniform quantization of weights in neural network

Jelena Nikolić,
Stefan Tomić,
Zoran Perić
et al.

Abstract: Motivated by the fact that optimizing quantization error fails to minimize the accuracy degradation caused by the quantization of Neural Network (NN) weights, in this paper, we highlight the need for a comprehensive analysis of post-training quantization, covering average bit rate, accuracy degradation and SQNR. One such analysis, for the layer-wise uniform quantization and its application is NN weight compression is presented in this paper. We introduce an additional aspect of Uniform Quantizer (UQ) choice, a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 23 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?