2022
DOI: 10.48550/arxiv.2202.08132
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Prospect Pruning: Finding Trainable Weights at Initialization using Meta-Gradients

Abstract: Pruning neural networks at initialization would enable us to find sparse models that retain the accuracy of the original network while consuming fewer computational resources for training and inference. However, current methods are insufficient to enable this optimization and lead to a large degradation in model performance. In this paper, we identify a fundamental limitation in the formulation of current methods, namely that their saliency criteria look at a single step at the start of training without taking… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 4 publications
0
4
0
Order By: Relevance
“…SNIP (Lee et al, 2018) is one of the pioneering works which aim to find trainable sub-networks without any training. Some following works (Wang et al, 2020a;Tanaka et al, 2020;de Jorge et al, 2020;Alizadeh et al, 2022) aim to propose different metrics to prune networks at initialization. Among them, Synflow (Tanaka et al, 2020), SPP (Lee et al, 2019), andFORCE (de Jorge et al, 2020) try to address the problem of layer collapse during pruning.…”
Section: Neural Network Pruningmentioning
confidence: 99%
“…SNIP (Lee et al, 2018) is one of the pioneering works which aim to find trainable sub-networks without any training. Some following works (Wang et al, 2020a;Tanaka et al, 2020;de Jorge et al, 2020;Alizadeh et al, 2022) aim to propose different metrics to prune networks at initialization. Among them, Synflow (Tanaka et al, 2020), SPP (Lee et al, 2019), andFORCE (de Jorge et al, 2020) try to address the problem of layer collapse during pruning.…”
Section: Neural Network Pruningmentioning
confidence: 99%
“…Finally, pruning on LeNet-5 and AlexNet can compress the number of parameters to 108 and 17.7 times, respectively, without loss of network performance. N. Lee and M. Alizadeh perform one-shot pruning before the network model is initialized [11,12]. Based on the importance of the weight connection, determine the importance of the weight and remove the weight with low importance.…”
Section: Unstructured Pruningmentioning
confidence: 99%
“…Criterion Method [8,9] weights weights magnitude train, prune and fine-tune [10] weights weights magnitude mask learning [11] weights weights magnitude prune and train [12] weights weights magnitude prune and train [14] filters L1 norm train, prune and fine-tune [15] filters filters magnitude group-LASSO regularization [16] filters magnitude of batchnorm parameters train, prune and fine-tune [17] filters output of the next layer train, prune and fine-tune [18] filters geometric median of common information in filters train, prune and fine-tune [19] filters average rank of feature map train, prune and fine-tune [20] filters channel independence train, prune and fine-tune [21,22] filters L p norm train, prune, and fine-tune…”
Section: Article Structurementioning
confidence: 99%
See 1 more Smart Citation