2019 26th IEEE International Conference on Electronics, Circuits and Systems (ICECS) 2019
DOI: 10.1109/icecs46596.2019.8964993
|View full text |Cite
|
Sign up to set email alerts
|

CoopNet: Cooperative Convolutional Neural Network for Low-Power MCUs

Abstract: Fixed-point quantization and binarization are two reduction methods adopted to deploy Convolutional Neural Networks (CNNs) on end-nodes powered by low-power microcontroller units (MCUs). While most of the existing works use them as stand-alone optimizations, this work aims at demonstrating there is margin for a joint cooperation that leads to inferential engines with lower latency and higher accuracy. Called CoopNet, the proposed heterogeneous model is conceived, implemented and tested on off-the-shelf MCUs wi… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(12 citation statements)
references
References 11 publications
0
12
0
Order By: Relevance
“…Furthermore, our hardware and software design provides a true real-time parallel inference scheme, which allows its users to exploit all resources for the entire time. Such customization of accelerators and their concurrent execution are big advantages of FPGAs, and this is fundamentally different from the sequential approaches that previous studies [18,19,20] took, which could make one network idle when the other network is running.…”
Section: Multiple Neural Network Architecturementioning
confidence: 88%
See 2 more Smart Citations
“…Furthermore, our hardware and software design provides a true real-time parallel inference scheme, which allows its users to exploit all resources for the entire time. Such customization of accelerators and their concurrent execution are big advantages of FPGAs, and this is fundamentally different from the sequential approaches that previous studies [18,19,20] took, which could make one network idle when the other network is running.…”
Section: Multiple Neural Network Architecturementioning
confidence: 88%
“…However, in this study, both DSP and LUT utilizations are maximized in a flexible and efficient way by implementing two different networks with different main computation units. This is an aspect that CPU-based solutions [18] or even previous FPGA-based solutions [19,20] did not offer. Previous FPGA-based solutions either implemented only one of the networks on FPGAs [20], which resulted in LUT being a bottleneck, or did not use extremely low bit-width networks [19], which resulted in DSP being a bottleneck.…”
Section: Multiple Neural Network Architecturementioning
confidence: 97%
See 1 more Smart Citation
“…Initially, elements of gradient-based learning were combined with other machine learning algorithms [12,19] to produce ultra-low memory classifiers (< 2KB). More recently, manually designed neural networks with quantisation and binarisation have been used for image classification on MCUs, too, though with a larger memory footprint [26,37]. Neural architecture search (NAS) is a widely explored topic in deep learning for GPUs.…”
Section: Related Workmentioning
confidence: 99%
“…There exist different works that proposed the use of ensemble methods for deep neural networks. Remarkable results are reported in [26], where the authors adopted a boosting strategy on image classification, but also in CoopNet [27] which combines multiple precision models to improve accuracy and inference latency. Even more interesting, the concept of ensemble learning can be found in the internal architecture of the most recent CNN models.…”
Section: Ensembles Learningmentioning
confidence: 99%