2021
DOI: 10.3390/electronics10222823
|View full text |Cite
|
Sign up to set email alerts
|

Integer-Only CNNs with 4 Bit Weights and Bit-Shift Quantization Scales at Full-Precision Accuracy

Abstract: Quantization of neural networks has been one of the most popular techniques to compress models for embedded (IoT) hardware platforms with highly constrained latency, storage, memory-bandwidth, and energy specifications. Limiting the number of bits per weight and activation has been the main focus in the literature. To avoid major degradation of accuracy, common quantization methods introduce additional scale factors to adapt the quantized values to the diverse data ranges, present in full-precision (floating-p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
4
2
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 33 publications
0
4
0
Order By: Relevance
“…To classified the cognitive training trace, three commonly EEG classification models used at present were compared: EEG-Net [33], 4-CNN [34], and EEG-GRUNet [35].…”
Section: Eeg Signal Classification Resultsmentioning
confidence: 99%
“…To classified the cognitive training trace, three commonly EEG classification models used at present were compared: EEG-Net [33], 4-CNN [34], and EEG-GRUNet [35].…”
Section: Eeg Signal Classification Resultsmentioning
confidence: 99%
“…To verify the validity of MTSC, three EEG classification models commonly used at present were compared: EEG-Net [28], 4-CNN [29], and EEG-GRUNet [30]. The experimental samples were input into the three classification models and MTSC.…”
Section: Eeg Signal Classification Resultsmentioning
confidence: 99%
“…2) Lower-bit Quantization: There are various quantization methods below 8 bits, including 4-bit quantization [118]- [120], 3-bit quantization [121], 2-bit quantization [122], [123], binary network [124]- [126], ternary network [127], [128]. These studies have achieved a very high compression ratio while retaining the performance of the original network as much as possible.…”
Section: B Quantization Of Different Bitsmentioning
confidence: 99%