2020
DOI: 10.1007/978-3-030-66770-2_22
|View full text |Cite
|
Sign up to set email alerts
|

Leveraging Automated Mixed-Low-Precision Quantization for Tiny Edge Microcontrollers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(16 citation statements)
references
References 9 publications
0
16
0
Order By: Relevance
“…However, uniformly quantizing a model to ultra low-precision can cause significant accuracy degradation. It is possible to address this with mixed-precision quantization [51,80,100,180,191,201,226,233,236,250,273]. In this approach, each layer is quantized with different bit precision, as illustrated in Figure 8.…”
Section: B Mixed-precision Quantizationmentioning
confidence: 99%
See 1 more Smart Citation
“…However, uniformly quantizing a model to ultra low-precision can cause significant accuracy degradation. It is possible to address this with mixed-precision quantization [51,80,100,180,191,201,226,233,236,250,273]. In this approach, each layer is quantized with different bit precision, as illustrated in Figure 8.…”
Section: B Mixed-precision Quantizationmentioning
confidence: 99%
“…Moreover, support for INT4 quantization was only recently added to TVM [254]. Recent work has shown that low precision and mixed-precision quantization with INT4/INT8 works in practice [51,80,100,106,180,191,201,226,233,233,236,250,254,273]. Thus, developing efficient software APIs for lower precision quantization will have an important impact.…”
Section: Future Directions For Research In Quantizationmentioning
confidence: 99%
“…Consequently, it is highly suggested to develop a new federated learning scheme to take into consideration the trade-off between utilization of communication resources and convergence rate throughout the training phase. Numerous quantization methods might be applied for federated learning across IoT networks including hyper-sphere quantization [78], low precision quantizer [79,80], and universal vector quantization [81].…”
Section: • Sparsification-empowered Federated Learningmentioning
confidence: 99%
“…Similarly, MCUNet (Lin et al, 2020) uses evolutionary search to design NNs for larger MCUs (2MB eFlash / 512KB SRAM) and larger datasets including visual wakewords (VWW) (Chowdhery et al, 2019) and keyword spotting (KWS) (Warden, 2018)). Reinforcement learning (RL) has also been used to choose quantization options in order to help fit an ImageNet model onto a larger MCU (2MB eFlash) (Rusci et al, 2020b). As well as images, audio tasks are an important driver for TinyML.…”
Section: Machine Learningmentioning
confidence: 99%
“…When compared to the sub-byte kernels of CMix-NN (Capotondi et al, 2020), our 4−bit kernels can substantially hide the latency overhead due to software-emulation of 4−bit operations, by fully exploiting the available instruction-level-parallelism (ILP) bandwidth on Cortex-M microcontrollers. Furthermore, we believe the accuracy of the 4−bit KWS MicroNet can be further improved by selectively quantizing lightweight depthwise layers to 8−bits, while quantizing remaining memory-and latency-heavy pointwise and standard convolutional layers to 4-bits (Rusci et al, 2020a;Gope et al, 2020). Table 2.…”
Section: Abstractpotting (Kws)mentioning
confidence: 99%