2019
DOI: 10.48550/arxiv.1911.07412
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Provable Filter Pruning for Efficient Neural Networks

Abstract: We present a provable, sampling-based approach for generating compact Convolutional Neural Networks (CNNs) by identifying and removing redundant filters from an over-parameterized network. Our algorithm uses a small batch of input data points to assign a saliency score to each filter and constructs an importance sampling distribution where filters that highly affect the output are sampled with correspondingly high probability. In contrast to existing filter pruning approaches, our method is simultaneously data… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
15
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 11 publications
(16 citation statements)
references
References 31 publications
1
15
0
Order By: Relevance
“…To validate the effectiveness of the proposed method on large-scale image classification datasets, we further apply GFBS on the ImageNet dataset and compare with more methods including [33,35,40]. We prune ResNet-18 and ResNet-50 and provide the results in Table 5.…”
Section: Results On Imagenetmentioning
confidence: 99%
“…To validate the effectiveness of the proposed method on large-scale image classification datasets, we further apply GFBS on the ImageNet dataset and compare with more methods including [33,35,40]. We prune ResNet-18 and ResNet-50 and provide the results in Table 5.…”
Section: Results On Imagenetmentioning
confidence: 99%
“…Discussion of other model compression techniques. We implemented our approach using two types of model compression techniques: (1) data-independent techniques: filter pruning [30] and Filter Pruning via Geometric Median (FPGM) [14] that directly calculate the filters' importance and remove unimportant ones; (2) data-dependent techniques: TaylorFOWeight [41], High-Rank Feature Map (HRank) [34], and Provable Filter Pruning (PFP) [32]) that calculate the filters' importance based on training samples. By testing ResNet18 and WideResNet as examples, Figure 14 shows that all techniques achieves similar accuracies under different available memories.…”
Section: Discussionmentioning
confidence: 99%
“…Sampling-based Pruning Recently, a number of works (Baykal et al, 2019a;Liebenwein et al, 2019;Baykal et al, 2019b;Mussay et al, 2020), proposed to procedures for pruning networks based on variants of (iterative) random sampling according to certain sensitivity score. These methods can provide concentration bounds on the difference of output between the pruned networks and the full networks, which may yield a bound of…”
Section: Related Workmentioning
confidence: 99%