2020
DOI: 10.1137/19m1246468
|View full text |Cite
|
Sign up to set email alerts
|

Fast Convex Pruning of Deep Neural Networks

Abstract: We develop a fast, tractable technique called Net-Trim for simplifying a trained neural network. The method is a convex post-processing module, which prunes (sparsifies) a trained network layer by layer, while preserving the internal responses. We present a comprehensive analysis of Net-Trim from both the algorithmic and sample complexity standpoints, centered on a fast, scalable convex optimization program. Our analysis includes consistency results between the initial and retrained models before and after Net… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
25
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 37 publications
(31 citation statements)
references
References 28 publications
1
25
0
Order By: Relevance
“…S. Ye et al [19] propose a progressive weight pruning approach and demonstrate high pruning rate by using partial pruning with moderate pruning rates. Aghasi et al [20] develop a convex post-processing technique that prunes a trained network layer by layer while preserving the internal responses.…”
Section: Background and Related Workmentioning
confidence: 99%
“…S. Ye et al [19] propose a progressive weight pruning approach and demonstrate high pruning rate by using partial pruning with moderate pruning rates. Aghasi et al [20] develop a convex post-processing technique that prunes a trained network layer by layer while preserving the internal responses.…”
Section: Background and Related Workmentioning
confidence: 99%
“…In this section, we first introduce the topology of the feedforward neural network models, then we explain the pruning method [37] that has been used to generate the compressed deep learning models. After that, we illustrate our approach to synthesis the compressed models, and the selection mechanism used to filter the best ones.…”
Section: Ensyth: Synthesis Of Deep Learning Ensemblesmentioning
confidence: 99%
“…The accuracy of LeNet-5 baseline model on CIFAR-10 is 78.4% , CIFAR-5 is 73.3% and 90.3% on MNIST-FASHION. After that, we prune and fine tune the baseline models with Net-Trim [37]. Net-trim's has four hyperparameters: L1: apply L1 regularisation on model's weight; L2: apply L2 regularisation on model's weight; dropout: a factor used to ignore neurons during a training process randomly; Epsilon gain: has a direct effect on the accuracy as well the sparsity of the pruned model.…”
Section: Network Training and Pruningmentioning
confidence: 99%
“…Jin et al [32] extended this method by restoring the pruned weights, training the network again, and repeating the process. Rather than pruning by thresholding, Aghasi et al [1,2] proposed Net-Trim, which prunes an already trained network layer by layer using convex optimization in order to ensure that the layer inputs and outputs remain consistent with the original network. For CNNs in particular, filter or channel pruning is preferred because it significantly reduces the amount of weight parameters required compared to individual weight pruning.…”
Section: Introductionmentioning
confidence: 99%