2020
DOI: 10.1109/tvlsi.2020.3002779
|View full text |Cite
|
Sign up to set email alerts
|

An Efficient Hardware Accelerator for Structured Sparse Convolutional Neural Networks on FPGAs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
68
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 93 publications
(68 citation statements)
references
References 30 publications
0
68
0
Order By: Relevance
“…Compared to the work in [26], this study uses the same frequency of 100 MHz, but our work achieved a performance improvement of 1.38× and energy efficiency by 1.63×. Compared to the accelerator in [36], this design uses only 471 DSPs and operates at a 100 MHz frequency, much less than those values in [36]. This work achieves a 2.64× improvement in sparsity, although its energy efficiency is lower than that in [36].…”
Section: B Experiments Results Of Fpga Implementationmentioning
confidence: 86%
See 1 more Smart Citation
“…Compared to the work in [26], this study uses the same frequency of 100 MHz, but our work achieved a performance improvement of 1.38× and energy efficiency by 1.63×. Compared to the accelerator in [36], this design uses only 471 DSPs and operates at a 100 MHz frequency, much less than those values in [36]. This work achieves a 2.64× improvement in sparsity, although its energy efficiency is lower than that in [36].…”
Section: B Experiments Results Of Fpga Implementationmentioning
confidence: 86%
“…Compared to the accelerator in [36], this design uses only 471 DSPs and operates at a 100 MHz frequency, much less than those values in [36]. This work achieves a 2.64× improvement in sparsity, although its energy efficiency is lower than that in [36]. Althought, the performance of the accelerator in [37] is higher when using 2 × the frequency and 2× the DSP resources, this work achieves higher energy efficiency.…”
Section: B Experiments Results Of Fpga Implementationmentioning
confidence: 97%
“…However, they only focused on the zeros in activation maps, while the parameters in filters' weights were not taken into considerations. On the other hand, there are several works [10] and [18] that have considered zeros in weights, but they are typically combined with zero compression methods to support sparse networks. In general, these methods require additional data buffers which would cause increased power consumption when the number of zeros is not so large or the target network is dense, which motivates us to develop a power-efficient accelerator design to support both dense and sparse networks.…”
Section: Preliminariesmentioning
confidence: 99%
“…Similar to Argus, SparseNN [17] and Cambricon-x [18] take advantage of skipping zeros in CNN weights. Beside mentioned, there are many other highquality architectures in terms of performance, like Eyeriss v2 [12], ENVISION [18], Thinker [19], UNPU [20], Snowflake [22], Caffeine [23], CoNNa [24], and architectures in [25]- [27].…”
Section: Introductionmentioning
confidence: 99%