2021
DOI: 10.3390/e23080933
|View full text |Cite
|
Sign up to set email alerts
|

Design of a 2-Bit Neural Network Quantizer for Laplacian Source

Abstract: Achieving real-time inference is one of the major issues in contemporary neural network applications, as complex algorithms are frequently being deployed to mobile devices that have constrained storage and computing power. Moving from a full-precision neural network model to a lower representation by applying quantization techniques is a popular approach to facilitate this issue. Here, we analyze in detail and design a 2-bit uniform quantization model for Laplacian source due to its significance in terms of im… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
29
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3

Relationship

3
4

Authors

Journals

citations
Cited by 10 publications
(29 citation statements)
references
References 30 publications
0
29
0
Order By: Relevance
“…Namely, NN normalization methods have already been confirmed to work well empirically [ 37 ], whereby a lot of them implicitly assume that the distributions of normalized NN parameters, primarily weights, initially have zero mean and unit variance. Keeping in mind that the weights’ distribution can closely fit some well-known probability density functions, such as the Laplacian probability density function (pdf) is [ 2 ], in this paper, as in [ 12 , 13 , 34 , 38 , 39 , 40 ], we assume the Laplacian-like distribution for experimental weights’ distribution and the Laplacian pdf for the theoretical distribution of weights to estimate the performance of the three-bit uniform quantizer (three-bit UQ) in question. Our motivation to address the simplest UQ also stems from the fact that UQs are not optimal for nonlinear distributions.…”
Section: Related Work and Motivationmentioning
confidence: 99%
See 3 more Smart Citations
“…Namely, NN normalization methods have already been confirmed to work well empirically [ 37 ], whereby a lot of them implicitly assume that the distributions of normalized NN parameters, primarily weights, initially have zero mean and unit variance. Keeping in mind that the weights’ distribution can closely fit some well-known probability density functions, such as the Laplacian probability density function (pdf) is [ 2 ], in this paper, as in [ 12 , 13 , 34 , 38 , 39 , 40 ], we assume the Laplacian-like distribution for experimental weights’ distribution and the Laplacian pdf for the theoretical distribution of weights to estimate the performance of the three-bit uniform quantizer (three-bit UQ) in question. Our motivation to address the simplest UQ also stems from the fact that UQs are not optimal for nonlinear distributions.…”
Section: Related Work and Motivationmentioning
confidence: 99%
“…For symmetrical data distributions, symmetrical quantizers with an even number of quantization steps are preferable [ 14 , 24 , 25 , 26 , 27 , 28 , 29 , 32 , 34 , 35 , 38 , 40 ]. Since it can be expected that most of the real data is asymmetrical, one could not conjecture that the preferable quantizer is also asymmetrical.…”
Section: Design Of Symmetric Three-bit Uniform Quantizer For the Purpose Of Nn Weights Compressionmentioning
confidence: 99%
See 2 more Smart Citations
“…It is well known that a nonuniform quantizer model, well accommodated to the signal's amplitude dynamic and a nonuniform pdf, has lower quantization error compared to the uniform quantizer (UQ) model with an equal number of quantization levels or equal bit-rates [2,11,13,18,[20][21][22][23][24][25][26][27]. However, due to the fact that UQ is the simplest quantizer model, it has been intensively studied, for instance in [23,24,[28][29][30][31][32].…”
Section: Introductionmentioning
confidence: 99%