2017 IEEE International Conference on Computer Vision (ICCV) 2017
DOI: 10.1109/iccv.2017.541
|View full text |Cite
|
Sign up to set email alerts
|

ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression

Abstract: We propose an efficient and unified framework, namely ThiNet, to simultaneously accelerate and compress CNN models in both training and inference stages. We focus on the filter level pruning, i.e., the whole filter would be discarded if it is less important. Our method does not change the original network structure, thus it can be perfectly supported by any off-the-shelf deep learning libraries. We formally establish filter pruning as an optimization problem, and reveal that we need to prune filters based on s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
1,206
1
7

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 1,635 publications
(1,219 citation statements)
references
References 22 publications
(38 reference statements)
5
1,206
1
7
Order By: Relevance
“…Interestingly, the discarded filters by our method are not only with small norms but may also have big values, which is significantly different from the results of the conventional filter pruning method, i.e. ThiNet [22]. Actually, the weights in filters for extracting color and texture information can be very small.…”
Section: Methodsmentioning
confidence: 64%
“…Interestingly, the discarded filters by our method are not only with small norms but may also have big values, which is significantly different from the results of the conventional filter pruning method, i.e. ThiNet [22]. Actually, the weights in filters for extracting color and texture information can be very small.…”
Section: Methodsmentioning
confidence: 64%
“…Therefore, in real-world applications, the developer can use ACNet to enhance a variety of models without exhausting parameter tunings, and the end-users can enjoy the performance improvement without slowing down the inference. Better still, since we introduce no custom structures into the deployed model, it can be future compressed via techniques including connection pruning [9,12], channel pruning [5,6,22,24], quantization [2,10,27], feature map compacting [34], etc.…”
Section: Architecture-neutral Cnn Structuresmentioning
confidence: 99%
“…An iterative pruning method which results in sparse networks is introduced in [14]. ThiNet [18] introduces a filter pruning technique which removes entire filters in convolutional neural networks. The authors propose a method using a least-squares approach to find channels corresponding to unimportant filters.…”
Section: Related Workmentioning
confidence: 99%