2023
DOI: 10.3390/electronics12051208
|View full text |Cite
|
Sign up to set email alerts
|

Deep Learning Architecture Improvement Based on Dynamic Pruning and Layer Fusion

Abstract: The heavy workload of current deep learning architectures significantly impedes the application of deep learning, especially on resource-constrained devices. Pruning has provided a promising solution to compressing the bloated deep learning models by removing the redundancies of the networks. However, existing pruning methods mainly focus on compressing the superfluous channels without considering layer-level redundancies, which results in the channel-pruned models still suffering from serious redundancies. To… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 47 publications
(45 reference statements)
0
2
0
Order By: Relevance
“…Pruning methods have been proposed to reduce the complexity of CNN models [13][14][15][16][17]. Channel pruning intends to exploit the redundancy of feature maps between channels and remove channels with the minimal performance loss [13].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Pruning methods have been proposed to reduce the complexity of CNN models [13][14][15][16][17]. Channel pruning intends to exploit the redundancy of feature maps between channels and remove channels with the minimal performance loss [13].…”
Section: Introductionmentioning
confidence: 99%
“…Channel pruning intends to exploit the redundancy of feature maps between channels and remove channels with the minimal performance loss [13]. Li et al [14] proposed pruning deep learning models using both channel-level and layer-level compression techniques. Liu et al [16] designed a pruning method that can be directly applied to existing modern CNN architectures by enforcing channel-level sparsity in the network to reduce the model size, decrease the run-time memory footprint and lower the number of computing operations while maintaining the accuracy of the model.…”
Section: Introductionmentioning
confidence: 99%