2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00198
|View full text |Cite
|
Sign up to set email alerts
|

Structured Compression by Weight Encryption for Unstructured Pruning and Quantization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4
1

Relationship

2
7

Authors

Journals

citations
Cited by 31 publications
(8 citation statements)
references
References 10 publications
0
8
0
Order By: Relevance
“…When parameters identified as unimportant are removed first, then quantization can be performed with a reduced number of parameters. Then, the number of quantization bits can also be reduced along with a smaller quantization error [51]. Both fine-grained pruning and structured pruning are investigated because of trade-offs between compression ratio and the regularity of memory access patterns [12], [52].…”
Section: Discussionmentioning
confidence: 99%
“…When parameters identified as unimportant are removed first, then quantization can be performed with a reduced number of parameters. Then, the number of quantization bits can also be reduced along with a smaller quantization error [51]. Both fine-grained pruning and structured pruning are investigated because of trade-offs between compression ratio and the regularity of memory access patterns [12], [52].…”
Section: Discussionmentioning
confidence: 99%
“…quantization [30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45], and through compression, i.e. sparsity/pruning [46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64], enables us to optimize the NN significantly. In quantization, low precision is used to represent the weights and activation, whereas in pruning the connections are completely removed.…”
Section: Reduced Precision and Compressionmentioning
confidence: 99%
“…Many methods are also proposed to determine weight-zero criteria, such as iterative threshold selection [39] and Hoffman code [40]. Kwon et al [41] proposed a sparsely quantized neural network weight representation scheme, specifically implemented by fine-grained and unstructured pruning methods.…”
Section: Unstructured Pruningmentioning
confidence: 99%