2017
DOI: 10.48550/arxiv.1702.04008
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Soft Weight-Sharing for Neural Network Compression

Abstract: The success of deep learning in numerous application domains created the desire to run and train them on mobile devices. This however, conflicts with their computationally, memory and energy intense nature, leading to a growing interest in compression. Recent work by Han et al. (2015a) propose a pipeline that involves retraining, pruning and quantization of neural network weights, obtaining state-of-the-art compression rates. In this paper, we show that competitive compression rates can be achieved by using a … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
76
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 58 publications
(77 citation statements)
references
References 20 publications
1
76
0
Order By: Relevance
“…LeNet-300-100 DNS [35] 1.99 1.79 L-OBS [11] 1.96 1.5 SWS [7] 1.94 4.3 Sparse VD [8] 1.92 1.47 Ours 1.98 ±0.07 1.51 ± 0.07 LeNet-5 DNS [35] 0.91 0.93 L-OBS [11] 1.66 0.9 SWS [7] 0.97 0.5 Sparse VD [8] 0.75 0.36 Ours 0.97 ± 0.05 0.65 ± 0.02 Table 1: Results for LeNet-300-100 and LeNet-5 trained and pruned on MNIST. For pruning, 1000 random training samples are chosen, α fc = 0.95, α conv = 0.9.…”
Section: Network Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…LeNet-300-100 DNS [35] 1.99 1.79 L-OBS [11] 1.96 1.5 SWS [7] 1.94 4.3 Sparse VD [8] 1.92 1.47 Ours 1.98 ±0.07 1.51 ± 0.07 LeNet-5 DNS [35] 0.91 0.93 L-OBS [11] 1.66 0.9 SWS [7] 0.97 0.5 Sparse VD [8] 0.75 0.36 Ours 0.97 ± 0.05 0.65 ± 0.02 Table 1: Results for LeNet-300-100 and LeNet-5 trained and pruned on MNIST. For pruning, 1000 random training samples are chosen, α fc = 0.95, α conv = 0.9.…”
Section: Network Methodsmentioning
confidence: 99%
“…Pruning assumes particular relevance for deep neural networks because modern architectures involve several millions of parameters. Existing pruning methods are based on different strategies, e.g Hessian analysis [2,3], magnitudes of weights [4], data-driven approaches [5,6], among others [7,8]. Pruning can be done in one shot [9] or in an iterative way [10], and it is possible to prune connections [2,3,11,12], neurons [13,14,6] or filters for convolutional layers [15,16].…”
Section: Introductionmentioning
confidence: 99%
“…We adopt the pruning technique for its simplicity, which seeks to induce sparse connections. There are many hybrid pruning methods [10,22,56] that are suitable for model deployment, but they may be overkill for our purpose of searching and designing the architecture after the compression. That being said, compression plays a completely different role in our work, namely it works as a tool for a better understanding of the underlying architecture and makes room for further improvements.…”
Section: Pruning-based Model Compressionmentioning
confidence: 99%
“…From a modelling perspective, specifying the functional forms of the prior and posterior distributions is an essential step to perform variational BNNs. One of the most commonly used variational family is fully factorized distribution referred to as the mean-field variational family (Graves, 2011;Blundell et al, 2015;Kingma et al, 2015;Neklyudov et al, 2017;Ullrich et al, 2017;Molchanov et al, 2017)…”
Section: Variational Bayesian Learning For Neural Networkmentioning
confidence: 99%
“…These, however, rely on prior and posterior pairs that are often chosen for convenience in inference, namely, computational tractability. The so-called meanfield variational family in these works assumes the posterior distributions to be all factorizing, and hence neglects the possibility of modelling statistical dependencies (i.e., correlations) among weight parameters (Graves, 2011;Blundell et al, 2015;Kingma et al, 2015;Neklyudov et al, 2017;Ullrich et al, 2017;Molchanov et al, 2017).…”
Section: Introductionmentioning
confidence: 99%