2020
DOI: 10.1609/aaai.v34i07.6910
|View full text |Cite
|
Sign up to set email alerts
|

Pruning from Scratch

Abstract: Network pruning is an important research field aiming at reducing computational costs of neural networks. Conventional approaches follow a fixed paradigm which first trains a large and redundant network, and then determines which units (e.g., channels) are less important and thus can be removed. In this work, we find that pre-training an over-parameterized model is not necessary for obtaining the target pruned structure. In fact, a fully-trained over-parameterized model will reduce the search space for the pru… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
71
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 133 publications
(71 citation statements)
references
References 13 publications
0
71
0
Order By: Relevance
“…(3) Unlike [7], only oneshot pruning is adopted by FL-PQSU, as the further pruning in federated training incurs additional overhead, but contributes little to performance improvement [13]. According to our testing experiments, 1 -norm based pruning outperforms blind model shrinking by about 1% accuracy loss for large models like VGG16, which can't be simply neglected [16].…”
Section: Structured Pruningmentioning
confidence: 99%
See 2 more Smart Citations
“…(3) Unlike [7], only oneshot pruning is adopted by FL-PQSU, as the further pruning in federated training incurs additional overhead, but contributes little to performance improvement [13]. According to our testing experiments, 1 -norm based pruning outperforms blind model shrinking by about 1% accuracy loss for large models like VGG16, which can't be simply neglected [16].…”
Section: Structured Pruningmentioning
confidence: 99%
“…Unlike these previous approaches, Play and Prune [15] allows to specify the error tolerance limit instead of the pruning ratio for each layer. Wang et al [16] verify that pruning from randomly initialized weights directly can result in more diverse pruned structures with competitive performance. More recent studies [7], [17] proposed to prune the model during training for not only improving the performance of inference but also reducing the costs of training.…”
Section: B Dnn Model Compressionmentioning
confidence: 99%
See 1 more Smart Citation
“…The implementation of our proposed compression technique 1 is assessed in terms of accuracy vs. parameter count against state-ofthe-art pruning approaches including [1,2,4,7,8,10,15,21,22,23]. We present strong results with our novel CNN compression applied to ResNet-50 [12] and MobileNet-V2 [13] architectures for image classification task on ImageNet ILSVRC dataset [24] (Sec.…”
Section: Fig 1 Tensor Reordering and Dct Compressionmentioning
confidence: 99%
“…Currently, there exist two main approaches to build lightweight networks. (1) The first approach uses network pruning techniques [ 9 ] or knowledge distillation [ 10 ] to achieve model compression and inference acceleration by removing redundant structures and parameters. Because the accuracy-focused models contain strategies that help overcome various problems encountered during training, such as overfitting, it is difficult to scale such a model down sufficiently without sacrificing accuracy.…”
Section: Introductionmentioning
confidence: 99%