2021
DOI: 10.48550/arxiv.2110.14032
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge

Abstract: Recently, a new trend of exploring sparsity for accelerating neural network training has emerged, embracing the paradigm of training on the edge. This paper proposes a novel Memory-Economic Sparse Training (MEST) framework targeting for accurate and fast execution on edge devices. The proposed MEST framework consists of enhancements by Elastic Mutation (EM) and Soft Memory Bound (&S) that ensure superior accuracy at high sparsity ratios. Different from the existing works for sparse training, this current work … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
1
1

Relationship

2
0

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 32 publications
0
2
0
Order By: Relevance
“…Pattern-based pruning [50,59,78] alleviates the shortcomings of prior works by incorporating the benefits from fine-grained pruning while maintaining structures that can be exploited for hardware accelerations with the help of compiler. Pattern-based pruning is a combination of kernel pattern pruning and connectivity pruning as shown in Fig.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Pattern-based pruning [50,59,78] alleviates the shortcomings of prior works by incorporating the benefits from fine-grained pruning while maintaining structures that can be exploited for hardware accelerations with the help of compiler. Pattern-based pruning is a combination of kernel pattern pruning and connectivity pruning as shown in Fig.…”
Section: Background and Related Workmentioning
confidence: 99%
“…To address this challenge, we envision that the lightweighted machine learning engine is needed to reduce the memory consumption for on-device training. For example, FedGKT [30], MEST [37], and TinyFL [38] are potential methods to reduce the training memory footprint for efficient on-device learning. To avoid catastrophic forgetting we also envision the use of clustering approaches that identify and store few core datasets from each time interval or leveraging IoT "Hubs" that can store non-sensitive/public datasets to inject memory in the training system.…”
Section: G Lifelong/continual Learningmentioning
confidence: 99%