Proceedings of the 2021 Great Lakes Symposium on VLSI 2021
DOI: 10.1145/3453688.3461739
|View full text |Cite
|
Sign up to set email alerts
|

Accommodating Transformer onto FPGA

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 27 publications
(6 citation statements)
references
References 13 publications
0
6
0
Order By: Relevance
“…As shown in Figure 1, despite the potential benefits of FPGA-based acceleration for transformer models, several challenges must be addressed to achieve efficient implementation [18][19][20]. A notable challenge is the long development cycle associated with hardware design, which may be exacerbated by the need to tune various high-level synthesis (HLS) parameters to optimize the final hardware architecture [14,18].…”
Section: Challenges In Fpga-based Transformer Accelerationmentioning
confidence: 99%
See 1 more Smart Citation
“…As shown in Figure 1, despite the potential benefits of FPGA-based acceleration for transformer models, several challenges must be addressed to achieve efficient implementation [18][19][20]. A notable challenge is the long development cycle associated with hardware design, which may be exacerbated by the need to tune various high-level synthesis (HLS) parameters to optimize the final hardware architecture [14,18].…”
Section: Challenges In Fpga-based Transformer Accelerationmentioning
confidence: 99%
“…A notable challenge is the long development cycle associated with hardware design, which may be exacerbated by the need to tune various high-level synthesis (HLS) parameters to optimize the final hardware architecture [14,18]. Various parameters, such as parallelism, memory type, and memory size, significantly affect FPGA resource utilization, performance, and power consumption [19,21,22]. In the realm of hardware design research, many studies focus on how to reduce model size through quantization methods, thereby shrinking the hardware footprint.…”
Section: Challenges In Fpga-based Transformer Accelerationmentioning
confidence: 99%
“…When comparing the characteristics implemented on a reconfigurable device with other works, to our knowledge, there is no research that has been able to realize a configurable accelerator, which can complete the whole-layer network inference and have compatible dense and sparse characteristics simultaneously. Authors of [5] and [12] show how to deploy a complete single-layer network, but the corresponding hardware for each operator on the dataflow is arranged, which is not configurable, and both of which are aimed at hardware acceleration for specific sparse matrices. Designs included in [8] and [9] have generic arrays for matrix operations.…”
Section: Characteristics Of Acceleratorsmentioning
confidence: 99%
“…Machine-Learning-as-a-Service (MLaaS) has emerged as a popular solution for accelerating inference in various applications [1]- [11]. The challenges of MLaaS comes from several folds: inference latency and privacy.…”
Section: Introductionmentioning
confidence: 99%