2019
DOI: 10.1109/access.2019.2947846
|View full text |Cite
|
Sign up to set email alerts
|

Compressing by Learning in a Low-Rank and Sparse Decomposition Form

Abstract: Kailing Guo and Xiaona Xie contributed equally to this work.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0
2

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 21 publications
0
5
0
2
Order By: Relevance
“…In [18], authors speed up convolutional layers by using depth-wise separable convolution and show a way to find the optimal channel configuration in groups to prevent a quality decrease. Modern methods also combine different approaches, for example, lowrank approximations and sparse filters, and reduce the number of parameters of modern architectures by about 30%, leaving accuracy almost at the same level [19].…”
Section: Related Workmentioning
confidence: 99%
“…In [18], authors speed up convolutional layers by using depth-wise separable convolution and show a way to find the optimal channel configuration in groups to prevent a quality decrease. Modern methods also combine different approaches, for example, lowrank approximations and sparse filters, and reduce the number of parameters of modern architectures by about 30%, leaving accuracy almost at the same level [19].…”
Section: Related Workmentioning
confidence: 99%
“…Sparse feature learning [14]- [19], L+S regularization of network weights [20] (U-Net trained with L+S output loss)…”
Section: Neural-model-based Mathematical-model-basedmentioning
confidence: 99%
“…There is plenty of room to incorporate sparsity-inducing priors into training as knowledge injection. They have explicitly been utilized for sparsification of hidden unit outputs in autoencoder-based sparse feature learning [14]- [19] as well as for low-rank and/or sparse regularization of network weights [5], [20]. The nuclear norm has not been used as a loss function yet, even though it is backpropable via automatic differentiation of the singular value decomposition [21].…”
Section: None (But Crucial To Prevent Catastrophic Forgetting)mentioning
confidence: 99%
“…For instance, network quantization methods have been proposed for storage space compression by decreasing the number of possible and unique values for the parameters [13,22,23]. Tensor decomposition approaches decompose network matrices into several smaller ones to estimate the informative parameters of the deep CNNs with low-rank approximation/factorization [24,25,26,27]. More recently, [28] also propose a framework of architecture distillation based on layer-wise replacement, called LightweightNet for memory and time saving.…”
Section: Related Workmentioning
confidence: 99%