2017
DOI: 10.1007/978-3-319-66709-6_15
|View full text |Cite
|
Sign up to set email alerts
|

Neuron Pruning for Compressing Deep Networks Using Maxout Architectures

Abstract: Abstract. This paper presents an efficient and robust approach for reducing the size of deep neural networks by pruning entire neurons. It exploits maxout units for combining neurons into more complex convex functions and it makes use of a local relevance measurement that ranks neurons according to their activation on the training set for pruning them. Additionally, a parameter reduction comparison between neuron and weight pruning is shown. It will be empirically shown that the proposed neuron pruning reduces… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(3 citation statements)
references
References 8 publications
(21 reference statements)
0
3
0
Order By: Relevance
“…It is essential to do a trade-off between the compressed rate and the accuracy reduction. Current popular methods include pruning [14] [15] [16], quantization [17], parameter sharing [18], knowledge distillation, lowrank approximation and direct design of compact models, etc. Besides, another similar method named Neural Architecture Search (NAS) is also widely considered to explore a suitable lightweight model for resource-limited devices.…”
Section: A Model Compressionmentioning
confidence: 99%
“…It is essential to do a trade-off between the compressed rate and the accuracy reduction. Current popular methods include pruning [14] [15] [16], quantization [17], parameter sharing [18], knowledge distillation, lowrank approximation and direct design of compact models, etc. Besides, another similar method named Neural Architecture Search (NAS) is also widely considered to explore a suitable lightweight model for resource-limited devices.…”
Section: A Model Compressionmentioning
confidence: 99%
“…Thus, it provides speedup and energy reduction; (2) Neuron pruning eliminates the entire rows/columns in the weight matrices reducing the weight matrices' dimensions proportionally, which could be efficiently implemented in the hardware compared to unstructured weight pruning [9]; (3) It also provides a way to determine the optimal number of neurons for a given network architecture [10]. Accordingly, many works have proposed various approaches to implement neuron pruning for the pursuit of a balance between compression ratio and accuracy [10][11][12][13][14].…”
Section: Introductionmentioning
confidence: 99%
“…In related work, [132] has applied AdaBoost to increase the resiliency of the overall system but has not explored in the energy reduction perspective. Moreover, the idea of DNS is completely different from neurons and weights pruning [133] performed during the training phase of the network to reduce the number of redundant neurons and parameters. Nevertheless, we can apply such methods to reduce each BL size.…”
Section: ) Approximationmentioning
confidence: 99%