2018
DOI: 10.3390/app8040504
|View full text |Cite
|
Sign up to set email alerts
|

An FPGA Implementation of a Convolutional Auto-Encoder

Abstract: Abstract:In order to simplify the hardware design and reduce the resource requirements, this paper proposes a novel implementation of a convolutional auto-encoder (CAE) in a field programmable gate array (FPGA). Instead of the traditional framework realized in a layer-by-layer way, we designed a new periodic layer-multiplexing framework for CAE. Only one layer is introduced and periodically reused to establish the network, which consumes fewer hardware resources. Moreover, by fixing the number of channels, thi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 25 publications
0
6
0
Order By: Relevance
“…The extracted features are global, local features are ignored, and local features are more important for wood texture recognition. Convolutional neural networks have the characteristics of local connection and weight sharing [35][36][37][38][39][40], which can accelerate the training of the network and facilitate the extraction of local features. The deep convolutional autoencoder designed in this paper is shown in Figure 2.…”
Section: Methods Of the Local Feature Descriptor Based On The Convolumentioning
confidence: 99%
“…The extracted features are global, local features are ignored, and local features are more important for wood texture recognition. Convolutional neural networks have the characteristics of local connection and weight sharing [35][36][37][38][39][40], which can accelerate the training of the network and facilitate the extraction of local features. The deep convolutional autoencoder designed in this paper is shown in Figure 2.…”
Section: Methods Of the Local Feature Descriptor Based On The Convolumentioning
confidence: 99%
“…In this Section, we compare our ternary AE with the most related state-of-the-art TNN implementations: TNN models for image classification implemented on customized accelerators in ASIC and FPGA [28], [38] and an AE model with 8-bit precision for image compression [39]. In [28] and [38], the TNNs are optimized for resourceefficiency and performance and use benchmarking datasets such as CIFAR100, SVHN, and GTSRB.…”
Section: Comparison With Existing Workmentioning
confidence: 99%
“…In our work, we calculate the computational roof and the I/O memory maximum bandwidth roof of the Xilinx ZYNQ 7100 computing platform according to Equation (10).…”
Section: The Roofline Model Of Zynq 7100mentioning
confidence: 99%
“…Because CNN is computationally intensive, it is not suitable for general-purpose processors, such as traditional CPUs. Many researchers have proposed CNN accelerators for implementation in the Field-programmable gate array (FPGA) [10,11], graphics processing unit (GPU) [3], and application-specific integrated circuit 2 of 18 (ASIC) [12]. These accelerators provide an order of magnitude performance improvement and energy advantage over general purpose processors [13].…”
Section: Introductionmentioning
confidence: 99%