2021
DOI: 10.1109/tcsii.2021.3095283
|View full text |Cite
|
Sign up to set email alerts
|

An FPGA-Based Energy-Efficient Reconfigurable Convolutional Neural Network Accelerator for Object Recognition Applications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
16
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 43 publications
(16 citation statements)
references
References 20 publications
0
16
0
Order By: Relevance
“…For other designs, the works [ 14 , 37 , 38 ] all used parameterizable configuration and data multiplexing, in which the authors of [ 14 ] optimised the instructions, used the ping-pong storage method, and [ 38 ] reduced the number of data accesses through kernel partitioning. The works of [ 38 , 39 ] implemented hardware using advanced chip technology on Intel FPGAs. The authors of [ 39 ] constructed a generic CNN compiler to generate customised FPGA hardware for different CNN inference tasks.…”
Section: Experimental Assessment and Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…For other designs, the works [ 14 , 37 , 38 ] all used parameterizable configuration and data multiplexing, in which the authors of [ 14 ] optimised the instructions, used the ping-pong storage method, and [ 38 ] reduced the number of data accesses through kernel partitioning. The works of [ 38 , 39 ] implemented hardware using advanced chip technology on Intel FPGAs. The authors of [ 39 ] constructed a generic CNN compiler to generate customised FPGA hardware for different CNN inference tasks.…”
Section: Experimental Assessment and Resultsmentioning
confidence: 99%
“…Simultaneously, the usage of the accelerator LUT was significantly reduced because the binary convolution computation uses fewer register resources. The accelerator performance at the convolutional layer was 35.66 GOPS/W, which was 1.11 times of [ 41 ] and 1.29 times of [ 38 ]. The overall neural network accelerator performance was 29.36 GOP/W, which was 2.07 times of [ 37 ], 1.37 times of [ 14 ], 1.21 times of [ 39 ] and 1.5 times of [ 40 ].…”
Section: Experimental Assessment and Resultsmentioning
confidence: 99%
“…Over the last few decades, FPGA boards have grown in prominence, particularly in the disciplines of machine vision and robotics [29,30]. FPGA technology has a number of advantages over software operating on a CPU or GPU, including faster execution [31] and lower power consumption [32]. The FPGA hardware, as well as a high-performance conventional server, might be housed in the cloud in the suggested scheme.…”
Section: Resultsmentioning
confidence: 99%
“…The AI CPs, such as Field Programmable Gate Array (FPGA), GPU, and Application Specific Integrated Circuit (ASIC), handle the AI workload and leave the rest of the task to the CPU [6,7]. FPGA has been treated as a promising solution to supplant conventional processors for performing computation-intensive tasks [8][9][10][11], such as deep neural network (DNN)-based image recognition, on the UAV platforms without violating size, weight, and power constraints inherent to UAV design [12][13][14][15]. Unless otherwise stated, we adopt FPGAs as the CPs in the following analysis.…”
Section: Introductionmentioning
confidence: 99%