Proceedings of the 50th Annual IEEE/ACM International Symposium on Microarchitecture 2017
DOI: 10.1145/3123939.3124552
|View full text |Cite
|
Sign up to set email alerts
|

C ir CNN

Abstract: Large-scale deep neural networks (DNNs) are both compute and memory intensive. As the size of DNNs continues to grow, it is critical to improve the energy efficiency and performance while maintaining accuracy. For DNNs, the model size is an important factor affecting performance, scalability and energy efficiency. Weight pruning achieves good compression ratios but suffers from three drawbacks: 1) the irregular network structure after pruning, which affects performance and throughput; 2) the increased training… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
12
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 179 publications
(12 citation statements)
references
References 70 publications
0
12
0
Order By: Relevance
“…We do an extensive comparison of HMD with 2 other compression techniques -model pruning and matrix factorization. Additionally, we also compared HMD with a structured matrix-based compression technique called block circular decomposition (BCD) [2,9]. BCD-compressed networks were able to recover the baseline accuracy for 2× -4× compression.…”
Section: Resultsmentioning
confidence: 99%
“…We do an extensive comparison of HMD with 2 other compression techniques -model pruning and matrix factorization. Additionally, we also compared HMD with a structured matrix-based compression technique called block circular decomposition (BCD) [2,9]. BCD-compressed networks were able to recover the baseline accuracy for 2× -4× compression.…”
Section: Resultsmentioning
confidence: 99%
“…CirCNN networks. CirCNN implementation of neural networks in introduced in [2] is a promising approach to reduce number of parameters while preserving networks' topologies. This is done by replacing matrices and convolution kernels in a neural network by block circulant matrices and block circulant convolution kernels.…”
Section: Notations and Preliminariesmentioning
confidence: 99%
“…For a convolution layer with kernel W = (W ijk ), where the k and indices represent the input and output channels and the i and j indices represent 2D kernels, we say W is block-circulant if for each fixed i, j the resulting matrix (W ijk ) k is block-circulant. As suggested in [2], for a single layer the block size should be a constant, while one can choose different block sizes for different layers. Using CirCNN implementation can significantly reduce number of learnable parameters.…”
Section: Notations and Preliminariesmentioning
confidence: 99%
See 1 more Smart Citation
“…Structured matrices have shown significant potential for compression of NN (Sindhwani et al, 2015;Ding et al, 2017;Cheng et al, 2015;Thakker et al, 2020). Block circular compression is an extension of structured matrix based compression technique, converting every block in a matrix into structured matrix.…”
Section: Related Workmentioning
confidence: 99%