2022
DOI: 10.1007/s10489-021-02802-8
|View full text |Cite
|
Sign up to set email alerts
|

Graph pruning for model compression

Abstract: Previous AutoML pruning works utilized individual layer features to automatically prune filters. We analyze the correlation for two layers from different blocks which have a short-cut structure. It is found that, in one block, the deeper layer has many redundant filters which can be represented by filters in the former layer so that it is necessary to take information from other layers into consideration in pruning. In this paper, a graph pruning approach is proposed, which views any deep model as a topology g… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 48 publications
0
6
0
Order By: Relevance
“…We refer to [54] for a more comprehensive NAS review. Block-wise weight-sharing NAS [36,46,83,84] approaches factorize the supernet into independently optimized blocks and thus reduce the weight-sharing space, resolving the issue of inaccurate architecture ratings caused by weightsharing. DNA [36] first introduced the block-wisely supervised architecture rating scheme with knowledge distillation.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…We refer to [54] for a more comprehensive NAS review. Block-wise weight-sharing NAS [36,46,83,84] approaches factorize the supernet into independently optimized blocks and thus reduce the weight-sharing space, resolving the issue of inaccurate architecture ratings caused by weightsharing. DNA [36] first introduced the block-wisely supervised architecture rating scheme with knowledge distillation.…”
Section: Related Workmentioning
confidence: 99%
“…Based on this scheme, DONNA [46] further propose to predict an architecture rating using a linear combination of its blockwise ratings rather than a simplistic sum. SP [83] were the first to apply this scheme to network pruning. However, all of the aforementioned methods rely on a supervised distillation scheme, which inevitably introduces architectural bias from the teacher.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…H. Zhu and Suyog Gupta showed comparison between the performance of large-sparse models and small dense model on a large variety of datasets and shows that large sparse model outperforms the former. Zhang, et al in their paper Graph Pruning for Model Compression discuss in filter pruning used in conjunction with graph convolution [3]. Filter pruning is a method in which selected filters are removed and a narrower model is rebuilt.…”
Section: Related Workmentioning
confidence: 99%

Model Compression

Ishtiaq,
Mahmood,
Anees
et al. 2021
Preprint