2020
DOI: 10.1007/978-3-030-61616-8_6
|View full text |Cite
|
Sign up to set email alerts
|

Pruning Artificial Neural Networks: A Way to Find Well-Generalizing, High-Entropy Sharp Minima

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

3
6

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 4 publications
0
5
0
Order By: Relevance
“…Furthermore, Han et al observed such a phenomenon for unstructured pruning [16]. Recently, it has been showed that the average entropy in the bottleneck layer for pruned backbones (in our case, x B ) is higher than in non-pruned ones [21]: this results in the propagation of less specific and more general information, which prevents features overfit, on top of which capsules layers can extract much more accurate information.…”
Section: Resultsmentioning
confidence: 61%
“…Furthermore, Han et al observed such a phenomenon for unstructured pruning [16]. Recently, it has been showed that the average entropy in the bottleneck layer for pruned backbones (in our case, x B ) is higher than in non-pruned ones [21]: this results in the propagation of less specific and more general information, which prevents features overfit, on top of which capsules layers can extract much more accurate information.…”
Section: Resultsmentioning
confidence: 61%
“…The lottery ticket hypothesis. It is a known fact that deep neural networks are typically over-parametrized and, after training, a part of the parameters can be removed without harming the performance, or even slightly improving the performance in small pruning regimes [7,13]. However, an interesting question rises: is it possible to train a subnetwork, still achieving the same performance as training the full model?…”
Section: The Lottery Of the Initializationmentioning
confidence: 99%
“…Many techniques have been proposed recently to reduce such complexity [15][16][17]. These approaches include the so-called pruning techniques, whose aim is to detect and remove the irrelevant parameters from a model [18].…”
Section: The European Distributed Deep Learning Librarymentioning
confidence: 99%