2020
DOI: 10.1145/3371157
|View full text |Cite
|
Sign up to set email alerts
|

3PXNet

Abstract: As the adoption of Neural Networks continues to proliferate different classes of applications and systems, edge devices have been left behind. Their strict energy and storage limitations make them unable to cope with the sizes of common network models. While many compression methods such as precision reduction and sparsity have been proposed to alleviate this, they don't go quite far enough. To push size reduction to its absolute limits, we combine binarization with sparsity in Pruned-Permuted-Packed XNOR Netw… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 41 publications
0
10
0
Order By: Relevance
“…Based on XNOR-Net, Ref. [18] constructed a pruned-permuted-packed network that combines binarization with sparsity to push model size reduction to very low limits. On the Nucleo platforms and Raspberry Pi, 3PXNet achieves a reduction in the model size by up to 38× and an improvement in runtime and energy of 25× compared to already compact conventional binarized implementations with a reduction in accuracy of less than 3%.…”
Section: Algorithmic Techniques For Low-power Edge Aimentioning
confidence: 99%
“…Based on XNOR-Net, Ref. [18] constructed a pruned-permuted-packed network that combines binarization with sparsity to push model size reduction to very low limits. On the Nucleo platforms and Raspberry Pi, 3PXNet achieves a reduction in the model size by up to 38× and an improvement in runtime and energy of 25× compared to already compact conventional binarized implementations with a reduction in accuracy of less than 3%.…”
Section: Algorithmic Techniques For Low-power Edge Aimentioning
confidence: 99%
“…Dedicated runtime libraries have also been developed, like ARM CMSIS-NN [43], TensorFlow Lite [21], or Microsoft EdgeML [29,41]. Pruning networks to high sparsity levels to save storage and computation has also been extensively explored [45,57].…”
Section: Accessibility Of Edge MLmentioning
confidence: 99%
“…Besides storage compression, binarization offers impressive performance improvements by replacing integer multiplication with bitwise XNOR operations [17,35,56]. Due to their efficiency, multiple software [34,50,54,66], and hardware [6,7,9,15,26,37,44,65] implementations of binarized neural networks have been proposed; as well as further optimizations, such as combining binarization with pruning [57], or memoization [48]. However, BNNs come with disadvantages, the most important of which is accuracy degradation [25,27].…”
Section: Accessibility Of Edge MLmentioning
confidence: 99%
See 2 more Smart Citations