2021
DOI: 10.1145/3476997
|View full text |Cite
|
Sign up to set email alerts
|

HW-FlowQ: A Multi-Abstraction Level HW-CNN Co-design Quantization Methodology

Abstract: Model compression through quantization is commonly applied to convolutional neural networks (CNNs) deployed on compute and memory-constrained embedded platforms. Different layers of the CNN can have varying degrees of numerical precision for both weights and activations, resulting in a large search space. Together with the hardware (HW) design space, the challenge of finding the globally optimal HW-CNN combination for a given application becomes daunting. To this end, we propose HW-FlowQ, a systematic approach… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
23
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 11 publications
(23 citation statements)
references
References 32 publications
0
23
0
Order By: Relevance
“…EMOMC [48] HW_FlowQ [49] APQ [47] This We have shown that the SRU-based model can be compressed using post-training quantization up to 8x without any error increase and up to 12x with a 1.2 percentage point (p.p.) increase in the error.…”
Section: Tablementioning
confidence: 77%
See 4 more Smart Citations
“…EMOMC [48] HW_FlowQ [49] APQ [47] This We have shown that the SRU-based model can be compressed using post-training quantization up to 8x without any error increase and up to 12x with a 1.2 percentage point (p.p.) increase in the error.…”
Section: Tablementioning
confidence: 77%
“…EMOMC (Evolutionary Multi-Objective Model Compression) [48], and HW_FlowQ [49] are methods that apply multi-objective evolutionary optimization to search for the NN model compression parameters. They have different approaches to overcoming the difficulty of applying retraining during the search.…”
Section: Group B: Multi-objective Hardware-aware Compressionmentioning
confidence: 99%
See 3 more Smart Citations