2019
DOI: 10.1109/jetcas.2019.2952137
|View full text |Cite
|
Sign up to set email alerts
|

Memory-Reduced Network Stacking for Edge-Level CNN Architecture With Structured Weight Pruning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 19 publications
(4 citation statements)
references
References 25 publications
0
4
0
Order By: Relevance
“…As shown in Table 1, three baseline DCNN architectures are used for evaluations of three different compression techniques described in Section II; quantization(T 0 ), pruning(T 1 ), and channel scaling(T 2 ), where we set f 0 = 10, f 1 = 10, and f 2 = 4, respectively. Based on the prior researches [16], [20], [21], the maximum compression levels from three different optimization approaches are achieved by adopting 8-bit quantization, 99% weight pruning, and 0.25-scaled channels.…”
Section: Resultsmentioning
confidence: 99%
“…As shown in Table 1, three baseline DCNN architectures are used for evaluations of three different compression techniques described in Section II; quantization(T 0 ), pruning(T 1 ), and channel scaling(T 2 ), where we set f 0 = 10, f 1 = 10, and f 2 = 4, respectively. Based on the prior researches [16], [20], [21], the maximum compression levels from three different optimization approaches are achieved by adopting 8-bit quantization, 99% weight pruning, and 0.25-scaled channels.…”
Section: Resultsmentioning
confidence: 99%
“…1/3 of weights are pruned for ResNet and AlexNet models when tested on CIFAR dataset with an accuracy drop of around 0.8% in 100 epochs. S. Moon et al [14] have proposed a novel memory-reduced multiple accuracy pruning method. This method is combination of multiple CNN optimization techniques.…”
Section: A Network Compression Using Pruning Methodsmentioning
confidence: 99%
“…Some studies ( Zhang et al, 2018 ; Moon et al, 2019 ) have focused on weight pruning, which is a common type of unstructured pruning and involves removing individual weights or neurons from the network without any constraints on their location or connectivity. Weight pruning is very effective in reducing the number of parameters and computations in a network, as it allows for fine-grained control over the sparsity level and can achieve very high compression ratios.…”
Section: Related Workmentioning
confidence: 99%