2021 IEEE International Conference on Image Processing (ICIP) 2021
DOI: 10.1109/icip42928.2021.9506708
|View full text |Cite
|
Sign up to set email alerts
|

On the Role of Structured Pruning for Neural Network Compression

Abstract: This works explores the benefits of structured parameter pruning in the framework of the MPEG standardization efforts for neural network compression. First less relevant parameters are pruned from the network, then remaining parameters are quantized and finally quantized parameters are entropy coded. We consider an unstructured pruning strategy that maximizes the number of pruned parameters at the price of randomly sparse tensors and a structured strategy that prunes fewer parameters yet yields regularly spars… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 8 publications
0
4
0
Order By: Relevance
“…Pruning In recent years, the focus of NNs pruning, shifted from unstructured (parameters are removed independently) to structured (entire neurons are zeroed-out). A recent work [18] even showed that structured procedures greatly benefit endto-end compression and the deployment on embedded devices. Following this trend, for our experiments, we used the structured pruning strategy described in [19].…”
Section: Methodsmentioning
confidence: 99%
“…Pruning In recent years, the focus of NNs pruning, shifted from unstructured (parameters are removed independently) to structured (entire neurons are zeroed-out). A recent work [18] even showed that structured procedures greatly benefit endto-end compression and the deployment on embedded devices. Following this trend, for our experiments, we used the structured pruning strategy described in [19].…”
Section: Methodsmentioning
confidence: 99%
“…As an effect, they minimize the cardinality of some i-th intermediate layer's output ∥x i ∥ 0 . Bragagnolo et al [20] showed that structured sparsity, despite removing significantly less parameters from the model, yields lower model's memory footprint and inference time. When pruning a network in a structured way, a simplification step which practically reduces the rank of the matrices is possible; on the other side, encoding unstructured sparse matrices lead to representation overheads [10].…”
Section: Effect Of Pruned Backbones To Capsule Layersmentioning
confidence: 99%
“…Can they be determined earlier, during a normal vanilla training? deployment time and making inference more efficient [7,8,9,10,11]. A recent work, the lottery ticket hypothesis [12], suggests that the fate of a parameter, namely whether it is useful for training (winner at the lottery of initialization) or if it can be removed from the architecture, is decided already at the initialization step.…”
Section: Vanilla Trainingmentioning
confidence: 99%
“…return M 11: end procedure where u i is some generic update term. In principle, the parameters in W are not in the model, and for instance they should not be included in the computation anymore; however, we still need to encode that are missing, producing an overhead, as they are removed in an unstructured way [9]. 2 Limits.…”
Section: The Lottery Of the Initializationmentioning
confidence: 99%