2020
DOI: 10.1109/access.2020.3023946
|View full text |Cite
|
Sign up to set email alerts
|

FPGAN: An FPGA Accelerator for Graph Attention Networks With Software and Hardware Co-Optimization

Abstract: The Graph Attention Networks (GATs) exhibit outstanding performance in multiple authoritative node classification benchmark tests (including transductive and inductive). The purpose of this research is to implement an FPGA-based accelerator called FPGAN for graph attention networks that achieves significant improvement on performance and energy efficiency without losing accuracy compared with PyTorch baseline. It eliminates the dependence on digital signal processors (DSPs) and large amounts of on-chip memory … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(8 citation statements)
references
References 22 publications
0
8
0
Order By: Relevance
“…FPGAN ( Yan, Tong & Zhi, 2020 ) is based on FPGA to accelerate the inference process of GAT. FPGAN designs a shift calculation unit for the intensive exp operation in GAT, which eliminates the dependence of computing performance on DSP, and uses an exponential approximation algorithm to fit SoftMax to normalize the attention coefficient.…”
Section: Fpga Based Hardware Acceleratorsmentioning
confidence: 99%
See 2 more Smart Citations
“…FPGAN ( Yan, Tong & Zhi, 2020 ) is based on FPGA to accelerate the inference process of GAT. FPGAN designs a shift calculation unit for the intensive exp operation in GAT, which eliminates the dependence of computing performance on DSP, and uses an exponential approximation algorithm to fit SoftMax to normalize the attention coefficient.…”
Section: Fpga Based Hardware Acceleratorsmentioning
confidence: 99%
“…To save memory and reduce the computational difficulty, FPGAN ( Yan, Tong & Zhi, 2020 ) compresses the model, and its core idea is to convert the weights to powers of 0 or 2 and judge whether retraining is required by observing the loss of accuracy after conversion. If the compression accuracy loss for one set is within a reasonable range, start the next set of compressions.…”
Section: Fpga Based Hardware Acceleratorsmentioning
confidence: 99%
See 1 more Smart Citation
“…2c, mean pool uses many ThCudaTensor_scatterAddKernels which are also present in the Aggregation phase. Again, similar to previous GNN accelerators [3], [4], [11], [21], [27], [39], [41]- [43], this work will focus only on the Aggregation and Combination phases, as the main kernels in aggregation and combination consume a majority of the GNN inference runtime.…”
Section: B Pytorch Geometric Characterizationmentioning
confidence: 99%
“…Computing GNN inference requires a mix of memory and compute intensive operations, which commodity CPUs, GPUs and traditional DNN accelerators do not exploit efficiently [27], [39], [40], [44]. This led to the development of many dedicated GNN accelerators, each with their own design methodology to extract as much performance as possible [3], [4], [11], [21], [27], [39], [41]- [43].…”
Section: Introductionmentioning
confidence: 99%