Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis 2019
DOI: 10.1145/3295500.3356156
|View full text |Cite
|
Sign up to set email alerts
|

PruneTrain

Abstract: State-of-the-art convolutional neural networks (CNNs) used in vision applications have large models with numerous weights. Training these models is very compute-and memory-resource intensive. Much research has been done on pruning or compressing these models to reduce the cost of inference, but little work has addressed the costs of training. We focus precisely on accelerating training. We propose PruneTrain, a cost-efficient mechanism that gradually reduces the training cost during training. PruneTrain uses a… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 51 publications
(10 citation statements)
references
References 40 publications
0
10
0
Order By: Relevance
“…• ElasticTrainer is time efficient. Compared to the existing schemes [28,34,64,67,73], it achieves up to 3.5× more training speedup in wall-clock time and reduces the training FLOPs by 60%.…”
Section: Offline Onlinementioning
confidence: 97%
See 2 more Smart Citations
“…• ElasticTrainer is time efficient. Compared to the existing schemes [28,34,64,67,73], it achieves up to 3.5× more training speedup in wall-clock time and reduces the training FLOPs by 60%.…”
Section: Offline Onlinementioning
confidence: 97%
“…A better choice is to adaptively adjust the trainable NN portion at runtime. NN pruning [43,65] for on-device training removes less important NN structures on the fly [50,64] (Figure 1 top-right). However, since the pruned NN portions can never be selected again even if they may be useful [52], NN's representation power is weakened over time and becomes insufficient for difficult learning tasks.…”
Section: Offline Onlinementioning
confidence: 99%
See 1 more Smart Citation
“…The IPT approach performs pruning and training step iteratively, and it is an iterative and greedy selection procedure to approximately optimize the non-convex problem for finding sparse structures in neural networks. Different from traditional greedy methods, like ThiNet (Luo, Wu, and Lin 2017) and PruneTrain (Lym et al 2019), that permanently cuts off the weights and will never be restored, IPT can reconstruct part of pruned weights to alleviate accuracy degradation. Sparse Evolutionary Training (SET) (Mocanu et al 2018) proposes a prune-regrowth procedure that allows the pruned neurons and connections to recover randomly.…”
Section: Iterative Prune-train Iptmentioning
confidence: 99%
“…The main tenet of QoA is that, in an end-to-end data processing system, one must consider various trade-offs of quality of data, processing time, cost, result accuracy, underlying computing capabilities, to name just a few, based on specific analysis context. Dealing with trade-offs in ML is one of the important research directions [32,33]. In our previous work, we also have examined QoA for common ML pipelines [34].…”
Section: Understanding the Quality Trade-offs In End-to-end Bim Objecmentioning
confidence: 99%