2018 Second International Conference on Advances in Electronics, Computers and Communications (ICAECC) 2018
DOI: 10.1109/icaecc.2018.8479517
|View full text |Cite
|
Sign up to set email alerts
|

An Efficient CNN Architecture for Image Classification on FPGA Accelerator

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
8
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(9 citation statements)
references
References 4 publications
1
8
0
Order By: Relevance
“…Table 8 compares our best architecture (Conv mixed-24) with existing works, which confirms that our architecture can substantially reduce hardware resources than the existing FGPA accelerators [ 28 , 33 , 34 , 35 ].…”
Section: Results and Analysissupporting
confidence: 61%
See 2 more Smart Citations
“…Table 8 compares our best architecture (Conv mixed-24) with existing works, which confirms that our architecture can substantially reduce hardware resources than the existing FGPA accelerators [ 28 , 33 , 34 , 35 ].…”
Section: Results and Analysissupporting
confidence: 61%
“…The energy consumption per image in the proposed accelerator is only 8.5 uJ, while it is 17.4 uJ in our previous accelerator [ 33 ]. Our energy per image is 1140, 81, and 555 times lower than the previous works [ 34 ], [ 28 ] and [ 35 ], respectively.…”
Section: Results and Analysismentioning
confidence: 54%
See 1 more Smart Citation
“…When the image resolution is higher, a common practice is to split the image into sub-blocks and use the sub-images for training the CNN model to perform defect inspection in metal AM [ 11 ]. The 200 × 200 image size used in our paper is comparable to literature with CPU-GPU approaches, and is considered the state-of-the-art comparing to other FPGA-based implementations (e.g., 28 × 28 in [ 16 ], 32 × 32 and 48 × 48 in [ 17 ]). Zhu et al [ 9 ] applied images with the size of 120 × 80 pixels to train the CNN for classification of weld surface defects.…”
Section: Research Methodsmentioning
confidence: 99%
“…Mujawar et al [ 16 ] proposed a 3-layer CNN architecture targeting at written digits recognition application on the MNIST dataset and implemented it in Artix-7 FPGAs. The authors also optimized the architecture by using loop-level parallel processing.…”
Section: Introductionmentioning
confidence: 99%