2020 4th International Conference on Intelligent Computing and Control Systems (ICICCS) 2020
DOI: 10.1109/iciccs48265.2020.9121004
|View full text |Cite
|
Sign up to set email alerts
|

Hardware Design of Approximate Matrix Multiplier based on FPGA in Verilog

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 10 publications
0
4
0
Order By: Relevance
“…Using blocking technology to improve bandwidth and storage efficiency, the ResNet-18 CNN model was implemented on XC7VX690T FPGA. Running at a clock frequency of 200MHz, the average throughput was 383 GOPS, which was a significant improvement in comparison to the work of Asgar Abbaszadeh et al In the same year, Ankit Gupta et al [81] made a tradeoff between accuracy and performance and proposed a new approximate matrix multiplier structure, which greatly improved the speed of matrix multiplication by introducing a negligible error amount and an approximate multiplication operation.…”
Section: ) Winograd Fast Convolution Algorithmmentioning
confidence: 96%
See 2 more Smart Citations
“…Using blocking technology to improve bandwidth and storage efficiency, the ResNet-18 CNN model was implemented on XC7VX690T FPGA. Running at a clock frequency of 200MHz, the average throughput was 383 GOPS, which was a significant improvement in comparison to the work of Asgar Abbaszadeh et al In the same year, Ankit Gupta et al [81] made a tradeoff between accuracy and performance and proposed a new approximate matrix multiplier structure, which greatly improved the speed of matrix multiplication by introducing a negligible error amount and an approximate multiplication operation.…”
Section: ) Winograd Fast Convolution Algorithmmentioning
confidence: 96%
“…As shown in Figure 6, according to different design concepts and requirements, FPGA-based neural network optimization technology can be roughly divided into optimization for data and operation, optimization for bandwidth, and optimization for memory and access, among others, which are introduced in detail below. [71][72][73][74][75][76][77][78], less computations [79][80][81], improve calculation speed [82][83][84][85], Winograd fast convolution algorithm [86][87][88][89][90][91], Im2col convolution optimization algorithm [92][93][94][95][96][97], pipelined design [98][99][100][101][102], Roofline model [103][104][105], ping-pong cache [106][107][108][109], input feature map reuse [110,111], filter reuse [111,112], convolutional reuse [110]…”
Section: Neural Network Optimization Technology Based On Fpgamentioning
confidence: 99%
See 1 more Smart Citation
“…hence, an aggressor can't figure out the design on account of the confusion, and the chip can't be overproduced with out ability of the key. moreover, design degree strategies comprehensive of cell disguise [8] will be utilized as equipment muddling and faker contacts are utilized to protect contrary to aggressors. The organization of in vogue cells with various functionalities is made to seem indistinguishable in the cover approach.…”
Section: Introductionmentioning
confidence: 99%