2017
DOI: 10.48550/arxiv.1705.08922
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Exploring the Regularity of Sparse Structure in Convolutional Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
66
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 82 publications
(66 citation statements)
references
References 11 publications
0
66
0
Order By: Relevance
“…This metric is directly related to the energy efficiency of (FPGA or ASIC) hardware implementation. As can be observed in the table, the proposed ADMM framework achieves significant amount of computation reduction compared with prior work, even when some [36,53] also focus on computation reductions. For the first metric of computation reduction, the improvement can be close to 3× compared with prior work for CONV layers, and this improvement reaches 3.6× for the second metric.…”
Section: Computation Reduction Comparisonsmentioning
confidence: 85%
See 1 more Smart Citation
“…This metric is directly related to the energy efficiency of (FPGA or ASIC) hardware implementation. As can be observed in the table, the proposed ADMM framework achieves significant amount of computation reduction compared with prior work, even when some [36,53] also focus on computation reductions. For the first metric of computation reduction, the improvement can be close to 3× compared with prior work for CONV layers, and this improvement reaches 3.6× for the second metric.…”
Section: Computation Reduction Comparisonsmentioning
confidence: 85%
“…Next we compare on the synthesized hardware speedup results between the proposed hardware-aware DNN model compression algorithm with baselines. The baselines include the iterative weight pruning and weight clustering work [22,24], and recent work [36,53]…”
Section: Results and Discussion On Computation Reduction And Hardware...mentioning
confidence: 99%
“…surg. [7] 80.0% +0.2% 3.1× NeST [5] 80.3% −0.1% 3.2× Fine-grained [30] Table 2 compares the pruning rates on the CONV layers vs. Top-5 accuracy, since the CONV layers are the most computationally intensive in state-of-art DNNs. We achieve 8.6× pruning in CONV layers with even slight accuracy enhancement, and 11.2× pruning with minor accuracy loss, consistently outperforming prior work in CONV layer weight pruning.…”
Section: Experimental Results On Weight Pruningmentioning
confidence: 99%
“…Sparsity in Deep Neural Nets has also been extensively explored, and the work in this area can be categorized as unstructured or structured types. In particular, unstructured pruning approaches (Han et al, 2015a,b;Dai et al, 2019;Mao et al, 2017;Narang et al, 2017a) result in random sparsity in weight matrices, which is difficult to accelerate on general-purpose hardware due to storage and memory access overheads. This has motivated structured sparsity-based approaches for both CNNs Wen et al, 2016;Narang et al, 2017a) and RNNs (Lu et al, 2016;Wen et al, 2017;Narang et al, 2017b).…”
Section: Related Workmentioning
confidence: 99%