2018 IEEE Statistical Signal Processing Workshop (SSP) 2018
DOI: 10.1109/ssp.2018.8450819
|View full text |Cite
|
Sign up to set email alerts
|

Simultaneous Sparsity and Parameter Tying for Deep Learning Using Ordered Weighted ℓ1 Regularization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 9 publications
0
1
0
Order By: Relevance
“…Similarly, in [42] proximal gradient descent with group OWL constraint (grOWL [43]) is used to simultaneously sparsify neurons and enforce parameter sharing. Proximal gradient descent is also used in [44] where Ordered Weighted 1 regularization (OWL [45]) allowing simultaneous sparsify weights and optimized weight sharing. In [46], filters in CNN layers are pruned by solving an optimization problem using a dedicated optimizer with either group LASSO or 2,0 regularization.…”
Section: Learning Structured Sparse Dnns Using Proximal Regularization Methodsmentioning
confidence: 99%
“…Similarly, in [42] proximal gradient descent with group OWL constraint (grOWL [43]) is used to simultaneously sparsify neurons and enforce parameter sharing. Proximal gradient descent is also used in [44] where Ordered Weighted 1 regularization (OWL [45]) allowing simultaneous sparsify weights and optimized weight sharing. In [46], filters in CNN layers are pruned by solving an optimization problem using a dedicated optimizer with either group LASSO or 2,0 regularization.…”
Section: Learning Structured Sparse Dnns Using Proximal Regularization Methodsmentioning
confidence: 99%