2022
DOI: 10.3390/s22155623
|View full text |Cite
|
Sign up to set email alerts
|

A Novel Deep-Learning Model Compression Based on Filter-Stripe Group Pruning and Its IoT Application

Abstract: Nowadays, there is a tradeoff between the deep-learning module-compression ratio and the module accuracy. In this paper, a strategy for refining the pruning quantification and weights based on neural network filters is proposed. Firstly, filters in the neural network were refined into strip-like filter strips. Then, the evaluation of the filter strips was used to refine the partial importance of the filter, cut off the unimportant filter strips and reorganize the remaining filter strips. Finally, the training … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 32 publications
0
2
0
Order By: Relevance
“…But the hierarchical model does not provide a structural view with any analytical framework; thus, a separate algorithm is integrated. Even in recent times, the researchers have directed the research model using a deep neural network where the design is completely based on complementary mobile terminals [13][14][15][16][17]. If each terminal in the mobile nodes is stationary, then insignifcant data flters will be removed from the system.…”
Section: Existing Approaches: a Surveymentioning
confidence: 99%
“…But the hierarchical model does not provide a structural view with any analytical framework; thus, a separate algorithm is integrated. Even in recent times, the researchers have directed the research model using a deep neural network where the design is completely based on complementary mobile terminals [13][14][15][16][17]. If each terminal in the mobile nodes is stationary, then insignifcant data flters will be removed from the system.…”
Section: Existing Approaches: a Surveymentioning
confidence: 99%
“…Pruning is one of the widely used compression techniques, which eliminates redundant connections or filters from the model to reduce its size without compromising performance [6]. However, sparse connectivity in the network resulting from pruning can lead to irregular memory access patterns and reduced performance during inference [10]. Implementing sparse networks on hardware, particularly on embedded devices with limited memory resources, can also present challenges.…”
mentioning
confidence: 99%