2021
DOI: 10.48550/arxiv.2107.07467
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Only Train Once: A One-Shot Neural Network Training And Pruning Framework

Abstract: Structured pruning is a commonly used technique in deploying deep neural networks (DNNs) onto resource-constrained devices. However, the existing pruning methods are usually heuristic, task-specified, and require an extra fine-tuning procedure. To overcome these limitations, we propose a framework that compresses DNNs into slimmer architectures with competitive performances and significant FLOPs reductions by Only-Train-Once (OTO). OTO contains two keys: (i) we partition the parameters of DNNs into zero-invari… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
10
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(10 citation statements)
references
References 65 publications
0
10
0
Order By: Relevance
“…Pruning is classified as either unstructured/weight pruning [14] or structured pruning [15]- [22] according to the method of determining the importance of parameters and removing them. Unstructured pruning [14] removes parameters by determining the importance of each parameter according to the saliency score of each algorithm.…”
Section: B Pruningmentioning
confidence: 99%
See 4 more Smart Citations
“…Pruning is classified as either unstructured/weight pruning [14] or structured pruning [15]- [22] according to the method of determining the importance of parameters and removing them. Unstructured pruning [14] removes parameters by determining the importance of each parameter according to the saliency score of each algorithm.…”
Section: B Pruningmentioning
confidence: 99%
“…Unstructured pruning also has the disadvantage of requiring specific library and hardware support for a sparse matrix. In contrast, structured pruning [15]- [22] judges the importance of network connections according to the saliency score of each algorithm based on larger units such as channels. Structured pruning does not have a sparse matrix so that an existing library can be used, and memory usage can be reduced without requiring additional hardware.…”
Section: B Pruningmentioning
confidence: 99%
See 3 more Smart Citations