2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) 2019
DOI: 10.1109/iccvw.2019.00245
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Structured Pruning and Architecture Searching for Group Convolution

Abstract: Efficient inference of Convolutional Neural Networks is a thriving topic recently. It is desirable to achieve the maximal test accuracy under given inference budget constraints when deploying a pre-trained model. Network pruning is a commonly used technique while it may produce irregular sparse models that can hardly gain actual speed-up. Group convolution is a promising pruning target due to its regular structure; however, incorporating such structure into the pruning procedure is challenging. It is because s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 33 publications
(75 reference statements)
0
7
0
Order By: Relevance
“…Weight pruning selects individual weights to be pruned. Because of the unstructured selection patterns, it requires customized GPU kernels or specialized hardware [17,58]. Filter pruning selects entire filters only.…”
Section: Dnn Compression Methodsmentioning
confidence: 99%
“…Weight pruning selects individual weights to be pruned. Because of the unstructured selection patterns, it requires customized GPU kernels or specialized hardware [17,58]. Filter pruning selects entire filters only.…”
Section: Dnn Compression Methodsmentioning
confidence: 99%
“…Although this method is flexible, it relies heavily on a learnable binary relationship matrix U, which introduces additional parameters that complicate the training of deep convolution neural networks. Ruizhe et al [24] formulated group-based convolution pruning as a channel permutation optimization problem and efficiently solved channel structural constraints using a heuristic algorithm. However, it was difficult to be determine the structural constraints of each convolutional layer.…”
Section: Related Workmentioning
confidence: 99%
“…To overcome this shortcoming, Zhuo et al [23] proposed a dynamic group convolution (DGC), in which each group is decided dynamically by a tiny auxiliary feature selector. Ruizhe et al [24] formulated group-based convolution pruning as a channel permutation optimization problem and solved channel structural constraints efficiently using a heuristic algorithm. However, it was difficult to determine the structural constraints of each convolutional layer.…”
Section: Introductionmentioning
confidence: 99%
“…Conceptually, model compression shares a similar formulation to NAS, i.e., the generalized formulation in Section 2.1 directly applies with either a regularization term for model complexity or a hard constraint for the maximal resource. Therefore, NAS approaches are often easily transferred for model compression [358], [417], including pruning [244], [359], [360], [364], [418], [419], quantization [224], [420], [421], [422], and joint optimization [423], [424], [425]. Sometimes, the searched configuration or connectivity mask 4.…”
Section: Model Compressionmentioning
confidence: 99%