2020
DOI: 10.1007/978-3-030-58574-7_27
|View full text |Cite
|
Sign up to set email alerts
|

HMQ: Hardware Friendly Mixed Precision Quantization Block for CNNs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
5

Relationship

1
9

Authors

Journals

citations
Cited by 45 publications
(11 citation statements)
references
References 24 publications
0
9
0
Order By: Relevance
“…One area uses uniform-precision quantization where the model shares the same precision Choukroun et al, 2019;Gong et al, 2019;Langroudi et al, 2019;Jin et al, 2020a;Bhalgat et al, 2020;Darvish Rouhani et al, 2020;Oh et al, 2021). Another direction studies mixed-precision that determines bit-width for each layer through search algorithms, aiming at better accuracy-efficiency trade-off (Dong et al, 2019;Wang et al, 2019;Habi et al, 2020;Fu et al, 2020;Yang & Jin, 2020;Zhao et al, 2021a;b;Ma et al, 2021b). There is also binarization network, which only applies 1-bit (Rastegari et al, 2016;Hubara et al, 2016;Cai et al, 2017;Bulat et al, 2020;.…”
Section: Related Workmentioning
confidence: 99%
“…One area uses uniform-precision quantization where the model shares the same precision Choukroun et al, 2019;Gong et al, 2019;Langroudi et al, 2019;Jin et al, 2020a;Bhalgat et al, 2020;Darvish Rouhani et al, 2020;Oh et al, 2021). Another direction studies mixed-precision that determines bit-width for each layer through search algorithms, aiming at better accuracy-efficiency trade-off (Dong et al, 2019;Wang et al, 2019;Habi et al, 2020;Fu et al, 2020;Yang & Jin, 2020;Zhao et al, 2021a;b;Ma et al, 2021b). There is also binarization network, which only applies 1-bit (Rastegari et al, 2016;Hubara et al, 2016;Cai et al, 2017;Bulat et al, 2020;.…”
Section: Related Workmentioning
confidence: 99%
“…Later, [2] suggests an end-to-end learning approach using a rate-distortion objective. To optimize performance under quantization, several works [16,21,52,54] use mixed-precision quantization, while others [9,18,28,32,39,40] propose post-quantization optimization techniques.…”
Section: Model Compressionmentioning
confidence: 99%
“…Power-of-Two Thresholds. A uniform, symmetric quantizer (either signed or unsigned) with a power-of-two integer threshold is said to be a hardwarefriendly quantizer [18]. Restricting the threshold of a symmetric quantizer to power-of-two integers (i.e.…”
Section: Background and Basic Notionsmentioning
confidence: 99%