2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) 2019
DOI: 10.1109/iccvw.2019.00306
|View full text |Cite
|
Sign up to set email alerts
|

Automated Multi-Stage Compression of Neural Networks

Abstract: The low-rank tensor approximation is very promising for the compression of deep neural networks. We propose a new simple and efficient iterative approach, which alternates low-rank factorization with a smart rank selection and fine-tuning. We demonstrate the efficiency of our method comparing to noniterative ones. Our approach improves the compression rate while maintaining the accuracy for a variety of tasks.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
36
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 45 publications
(46 citation statements)
references
References 19 publications
1
36
0
Order By: Relevance
“…It is mentioned in[46] that the difference among different random matrices is negligible. We have also confirmed this in our simulations 5. Depending on which side the random matrix is multiplied.VOLUME , 2020…”
supporting
confidence: 82%
See 2 more Smart Citations
“…It is mentioned in[46] that the difference among different random matrices is negligible. We have also confirmed this in our simulations 5. Depending on which side the random matrix is multiplied.VOLUME , 2020…”
supporting
confidence: 82%
“…This procedure is performed by multiplying a given matrix by a random matrix from the right-hand or left-hand side. It has been shown that this preserves the Euclidean distances among columns or rows approximately 5 [44], [45]. Let X ∈ R I×J be a given data matrix, and R be a target rank.…”
Section: A Random Projectionmentioning
confidence: 99%
See 1 more Smart Citation
“…Low-rank decomposition: Low-rank decomposition algorithms [ 30 , 31 , 32 ] use a lower-rank set instead of the original set of parameters to approximate the CNN to achieve compression. Swaminathan et al [ 31 ] argue that the low-rank decomposition of weight matrices should consider influence of both input as well as output neurons of a layer.…”
Section: Related Workmentioning
confidence: 99%
“…Additionally, low-rank approximations are also useful for speeding up the evaluation of convolutional neural networks [151] by using a low-rank representation of the filters, which are used for detecting image features. For a very similar task the authors in [121,122,172] rely on optimized tensor decompositions. Note that recently these techniques have also been applied to adversarial networks [48].…”
Section: Numerical Linear Algebra In Deep Learningmentioning
confidence: 99%