Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Towards Principled Graph Transformers

About

Graph learning architectures based on the k-dimensional Weisfeiler-Leman (k-WL) hierarchy offer a theoretically well-understood expressive power. However, such architectures often fail to deliver solid predictive performance on real-world tasks, limiting their practical impact. In contrast, global attention-based models such as graph transformers demonstrate strong performance in practice, but comparing their expressive power with the k-WL hierarchy remains challenging, particularly since these architectures rely on positional or structural encodings for their expressivity and predictive performance. To address this, we show that the recently proposed Edge Transformer, a global attention model operating on node pairs instead of nodes, has at least 3-WL expressive power. Empirically, we demonstrate that the Edge Transformer surpasses other theoretically aligned architectures regarding predictive performance while not relying on positional or structural encodings. Our code is available at https://github.com/luis-mueller/towards-principled-gts

Luis M\"uller, Daniel Kusuma, Blai Bonet, Christopher Morris• 2024

Related benchmarks

TaskDatasetResultRank
Graph RegressionZINC 12K (test)
MAE0.059
164
Graph RegressionZINC 250K (test)
MAE0.024
13
Algorithmic ReasoningCLRS-30 n=64 (test)
Sort Accuracy82.26
6
Showing 3 of 3 rows

Other info

Follow for update