2021
DOI: 10.48550/arxiv.2106.03893
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Rethinking Graph Transformers with Spectral Attention

Abstract: In recent years, the Transformer architecture has proven to be very successful in sequence processing, but its application to other data structures, such as graphs, has remained limited due to the difficulty of properly defining positions. Here, we present the Spectral Attention Network (SAN), which uses a learned positional encoding (LPE) that can take advantage of the full Laplacian spectrum to learn the position of each node in a given graph. This LPE is then added to the node features of the graph and pass… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
31
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(32 citation statements)
references
References 21 publications
1
31
0
Order By: Relevance
“…There are many attempts of leveraging transformer into the graph domain. Existing methods (Veličković et al, 2017;Kreuzer et al, 2021;Zhang et al, 2020;Ying et al, 2021) have improved the transformer architecture to fit the graph input by improving the attention map or replacing the positional embedding to fit the graph, or both, which are more related to our GSA. Methods like Graph Attention Networks (GAT) (Veličković et al, 2017) and Graph Transformer (GT) constrain the self-attention mechanism to neighboring nodes.…”
Section: Transformer For Graphmentioning
confidence: 99%
See 3 more Smart Citations
“…There are many attempts of leveraging transformer into the graph domain. Existing methods (Veličković et al, 2017;Kreuzer et al, 2021;Zhang et al, 2020;Ying et al, 2021) have improved the transformer architecture to fit the graph input by improving the attention map or replacing the positional embedding to fit the graph, or both, which are more related to our GSA. Methods like Graph Attention Networks (GAT) (Veličković et al, 2017) and Graph Transformer (GT) constrain the self-attention mechanism to neighboring nodes.…”
Section: Transformer For Graphmentioning
confidence: 99%
“…They surpass GNN baseline methods on the graph representation task. Spectral Attention Network (SAN) (Kreuzer et al, 2021) employs a learned positional encoding (LPE) of Laplacian spectrum to learn the position of nodes in a graph. Graph-BERT (Zhang et al, 2020) uses several types of relative positional encodings to embed the information about the edges within a subgraph.…”
Section: Transformer For Graphmentioning
confidence: 99%
See 2 more Smart Citations
“…The GNN that are used by state-of-the-art device placement methods mostly follow the message-passing paradigm, which is known to have inherent limitations. For example, the expressiveness of such GNN is bounded by the Weisfeiler-Lehman isomorphism hierarchy [23]. Also, GNNs are known to suffer from over-squashing [24], where there is a distortion of information propagation between distant nodes.…”
Section: A Challenges In Device Placementmentioning
confidence: 99%