2020
DOI: 10.48550/arxiv.2006.00896
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pruning via Iterative Ranking of Sensitivity Statistics

Abstract: With the introduction of SNIP [40], it has been demonstrated that modern neural networks can effectively be pruned before training. Yet, its sensitivity criterion has since been criticized for not propagating training signal properly or even disconnecting layers. As a remedy, GraSP [70] was introduced, compromising on simplicity. However, in this work we show that by applying the sensitivity criterion iteratively in smaller steps -still before training -we can improve its performance without difficult implemen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
12
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(13 citation statements)
references
References 42 publications
0
12
0
Order By: Relevance
“…3). Indeed, very recent work empirically confirms that iteration improves the performance of SNIP [46].…”
Section: Synflow Random Graspmentioning
confidence: 89%
“…3). Indeed, very recent work empirically confirms that iteration improves the performance of SNIP [46].…”
Section: Synflow Random Graspmentioning
confidence: 89%
“…Therefore, more recent works have focused on scoring weights without the need for training using first order (Lee et al, 2019;Tanaka et al, 2020) and second order (Wang et al, 2020;Lubana & Dick, 2021) information. Note that the pruning process can be applied in one-shot or iteratively (de Jorge et al, 2021;Verdenius et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…On the contrary, if too many parameters are removed at one time, almost all models suffer from accuracy drops. This finding makes a connection to the success of the iterative magnitude pruning [10,53,5,6,63], where usually a pruning process with a small pruning rate (e.g., 0.2) needs to be iteratively repeated for good performance.…”
Section: Introductionmentioning
confidence: 99%