2021
DOI: 10.1109/tcsvt.2020.2996231
|View full text |Cite
|
Sign up to set email alerts
|

Model Compression Using Progressive Channel Pruning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 54 publications
(12 citation statements)
references
References 22 publications
0
12
0
Order By: Relevance
“…Various retraining-based methods are proposed in the literature (Ström 1997;Han et al 2015;Molchanov, Ashukha, and Vetrov 2017;Frankle and Carbin 2019;Renda, Frankle, and Carbin 2020;Narang et al 2017;Guo et al 2020;Guo, Xu, and Ouyang 2023;Huang et al 2023) to improve the accuracy. Although some of them may show promise, they often require extensive training or tuning of hyper-parameters, leading to inefficiency and slow convergence.…”
Section: Related Workmentioning
confidence: 99%
“…Various retraining-based methods are proposed in the literature (Ström 1997;Han et al 2015;Molchanov, Ashukha, and Vetrov 2017;Frankle and Carbin 2019;Renda, Frankle, and Carbin 2020;Narang et al 2017;Guo et al 2020;Guo, Xu, and Ouyang 2023;Huang et al 2023) to improve the accuracy. Although some of them may show promise, they often require extensive training or tuning of hyper-parameters, leading to inefficiency and slow convergence.…”
Section: Related Workmentioning
confidence: 99%
“…Model compression can reduce computation cost and storage consumption, meanwhile maintaining the loss of performance within a small scale. Guo et al [27] proposed PCP (Progressive Channel Pruning), a model pruning method with automatically determined network structure, to prune channels iteratively. Recently, attention mechanisms have been inserted into the model pruning method for improving performance.…”
Section: B Comparison Experimentsmentioning
confidence: 99%
“…For a given neural network, model compression may be another effective method to reduce the parameters and computation. Researchers have proposed some model pruning techniques to reduce the model size with an allowed accuracy range by removing useless channels for easier acceleration or unimportant connections between neurons [36,37]. In deep neural networks, weights are usually stored in the form of 32-bit floating-point numbers [38], and the model quantization compresses the original network by reducing the number of bits required to represent each weight [28,39].…”
Section: Related Workmentioning
confidence: 99%