2019
DOI: 10.18359/rcin.4194
|View full text |Cite
|
Sign up to set email alerts
|

A Hardware Accelerator for the Inference of a Convolutional Neural network

Abstract: Convolutional Neural Networks (CNNs) are becoming increasingly popular in deep learning applications, e.g. image classification, speech recognition, medicine, to name a few. However, the CNN inference is computationally intensive and demanding a large among of memory resources. In this work is proposed a CNN inference hardware accelerator, which was implemented in a co-processing scheme. The aim is to reduce the hardware resources and achieve the better possible throughput. The design was implemented i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
7
0

Year Published

2019
2019
2025
2025

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 8 publications
(9 citation statements)
references
References 15 publications
0
7
0
Order By: Relevance
“…This consumption is taken by Vivado's power report of the implemented design and consisting of 94 mW static and 534 mW dynamic power. This is nearly 67 % lower than the other LeNet CNN architectures which are around 1800 mW [16] [17]. In the experimental setup, the images are loaded using the serial interface of the board and the result is shown on the LEDs of the board.…”
Section: Discussionmentioning
confidence: 89%
See 2 more Smart Citations
“…This consumption is taken by Vivado's power report of the implemented design and consisting of 94 mW static and 534 mW dynamic power. This is nearly 67 % lower than the other LeNet CNN architectures which are around 1800 mW [16] [17]. In the experimental setup, the images are loaded using the serial interface of the board and the result is shown on the LEDs of the board.…”
Section: Discussionmentioning
confidence: 89%
“…In other words, the output of the Python and FPGA designs give exactly the same result in each stage of the CNN. Moreover, for a fair comparison, the proposed accelerator is compared with the other LeNet CNN implementations in the literature having the same number of convolutional and fully connected layers [17] [24] [25]. The design of [24] is using Zynq Ultrascale FPGA and HLS is used in the development stage.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Various CNN implementations on FPGA have been reported in literature [11]- [13], focusing on different aspects, e.g., the optimization of only the convolutional layers [14], [15] or the overall accelerator throughput [16]. There are also SW/HW co-design solutions that exploit the aggregate power of both an embedded processor and the programmable logic [17], [18]. Some [19], [20] present resource-intensive designs and achieve high throughput disregarding power consumption, while others implement binary neural networks (BNNs) that achieve high power efficiency at the cost of reduced accuracy using binary weights/biases [21].…”
Section: Background and Literature Reviewmentioning
confidence: 99%
“…AI applications include prediction, recommendation, classification and recognition, object detection, natural language processing, autonomous systems, among others. The topics of the articles in this special issue include deep learning applied to medicine [1,3], support vector machines applied to ecosystems [2], human-robot interaction [4], clustering in the identification of anomalous patterns in communication networks [5], expert systems for the simulation of natural disaster scenarios [6], real-time algorithms of artificial intelligence [7], and big data analytics for natural disasters [8].…”
mentioning
confidence: 99%