2021
DOI: 10.3390/electronics10202514
|View full text |Cite
|
Sign up to set email alerts
|

A Multi-Cache System for On-Chip Memory Optimization in FPGA-Based CNN Accelerators

Abstract: In recent years, FPGAs have demonstrated remarkable performance and contained power consumption for the on-the-edge inference of Convolutional Neural Networks. One of the main challenges in implementing this class of algorithms on board an FPGA is resource management, especially with regard to memory. This work presents a multi-cache system that allows for noticeably shrinking the required on-chip memory with a negligible variation of timing performance and power consumption. The presented methods have been ap… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 46 publications
0
1
0
Order By: Relevance
“…Nonetheless, HLS tools focus on the mapping and allocation of low-level primitive operations, exploring a large design space that is partially affected by the specific application domain [34]. On the other hand, Register Transfer Level (RTL) designs are often used to achieve extremely optimized architectures for those applications with a highly structured nature, such as CNN workloads [34]- [36]. A common approach is to merge model compression and architecture development into an end-to-end design flow to take advantage of both these techniques [37].…”
Section: Introductionmentioning
confidence: 99%
“…Nonetheless, HLS tools focus on the mapping and allocation of low-level primitive operations, exploring a large design space that is partially affected by the specific application domain [34]. On the other hand, Register Transfer Level (RTL) designs are often used to achieve extremely optimized architectures for those applications with a highly structured nature, such as CNN workloads [34]- [36]. A common approach is to merge model compression and architecture development into an end-to-end design flow to take advantage of both these techniques [37].…”
Section: Introductionmentioning
confidence: 99%