2019 29th International Conference on Field Programmable Logic and Applications (FPL) 2019
DOI: 10.1109/fpl.2019.00036
|View full text |Cite
|
Sign up to set email alerts
|

FPGA-Based Training Accelerator Utilizing Sparseness of Convolutional Neural Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
11
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(11 citation statements)
references
References 9 publications
0
11
0
Order By: Relevance
“…A sparse CNN training accelerator was designed on VCU1525. The accelerator was implemented on a pre-trained CNN model with 85% parameters pruned [20]. However, these existing works mainly focused on cloud-level devices with abundant computation and memory resources.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…A sparse CNN training accelerator was designed on VCU1525. The accelerator was implemented on a pre-trained CNN model with 85% parameters pruned [20]. However, these existing works mainly focused on cloud-level devices with abundant computation and memory resources.…”
Section: Related Workmentioning
confidence: 99%
“…), the on-chip memory of an edge FPGA is not big enough to hold weights or features in every Conv layer. Therefore, several works [4,18,20] applied quantization or pruning to reduce of-chip memory access. However, unlike inference where compressed networks cause little accuracy decrease [7], these training works have not proved that their compression techniques can remain high accuracy on large datasets with dense networks.…”
Section: Andmentioning
confidence: 99%
“…Sparse Accelerators. Sparse accelerators [84,129,[168][169][170][171][172][173][174][175][176] address the inefficiencies caused by zeros contained in sparse matrices, which is a fundamentally different problem than padding introduced by transpose and dilated convolutions. EcoFlow can be incorporated to these accelerators to obtain aggregated benefits.…”
Section: Related Workmentioning
confidence: 99%
“…To aid the development of deep learning models on FPGAs, (Venieris and Bouganis, 2016) propose a framework for mapping CNNs on FPGAs. Furthermore, the authors (Ma et al, 2019;Nakahara et al, 2019) propose an FPGA base accelerator to leverage the sources of parallelism in order to achieve an efficient implementation of a deep convolutional neural network. Finally, (Zhu et al, 2020) presents a reconfigurable framework for training CNNs.…”
Section: Related Workmentioning
confidence: 99%