2022
DOI: 10.1109/tnnls.2021.3084527
|View full text |Cite
|
Sign up to set email alerts
|

SeReNe: Sensitivity-Based Regularization of Neurons for Structured Sparsity in Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 14 publications
(6 citation statements)
references
References 12 publications
0
6
0
Order By: Relevance
“…From this, we observe that reducing the complexity of the backbone would result in the overall reduction of the complexity for the entire model N , and towards this end pruning has already proved to be an effective approach [9,10]. Pruning approaches can be divided into two groups.…”
Section: Effect Of Pruned Backbones To Capsule Layersmentioning
confidence: 92%
See 2 more Smart Citations
“…From this, we observe that reducing the complexity of the backbone would result in the overall reduction of the complexity for the entire model N , and towards this end pruning has already proved to be an effective approach [9,10]. Pruning approaches can be divided into two groups.…”
Section: Effect Of Pruned Backbones To Capsule Layersmentioning
confidence: 92%
“…Unstructured pruning methods aim at minimizing the cardinality ∥w∥ 0 of the parameters in the model, regardless the output topology [16,17,18]. On the other hand, structured approaches drop groups of weight connections entirely, such as entire channels or filters, imposing a regular pruned topology [9,19]. As an effect, they minimize the cardinality of some i-th intermediate layer's output ∥x i ∥ 0 .…”
Section: Effect Of Pruned Backbones To Capsule Layersmentioning
confidence: 99%
See 1 more Smart Citation
“…A recent work [18] even showed that structured procedures greatly benefit endto-end compression and the deployment on embedded devices. Following this trend, for our experiments, we used the structured pruning strategy described in [19].…”
Section: Methodsmentioning
confidence: 99%
“…Can they be determined earlier, during a normal vanilla training? deployment time and making inference more efficient [7,8,9,10,11]. A recent work, the lottery ticket hypothesis [12], suggests that the fate of a parameter, namely whether it is useful for training (winner at the lottery of initialization) or if it can be removed from the architecture, is decided already at the initialization step.…”
Section: Vanilla Trainingmentioning
confidence: 99%