Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Where Did the Gap Go? Reassessing the Long-Range Graph Benchmark

About

The recent Long-Range Graph Benchmark (LRGB, Dwivedi et al. 2022) introduced a set of graph learning tasks strongly dependent on long-range interaction between vertices. Empirical evidence suggests that on these tasks Graph Transformers significantly outperform Message Passing GNNs (MPGNNs). In this paper, we carefully reevaluate multiple MPGNN baselines as well as the Graph Transformer GPS (Ramp\'a\v{s}ek et al. 2022) on LRGB. Through a rigorous empirical analysis, we demonstrate that the reported performance gap is overestimated due to suboptimal hyperparameter choices. It is noteworthy that across multiple datasets the performance gap completely vanishes after basic hyperparameter optimization. In addition, we discuss the impact of lacking feature normalization for LRGB's vision datasets and highlight a spurious implementation of LRGB's link prediction metric. The principal aim of our paper is to establish a higher standard of empirical rigor within the graph machine learning community.

Jan T\"onshoff, Martin Ritzert, Eran Rosenbluth, Martin Grohe• 2023

Related benchmarks

TaskDatasetResultRank
Graph ClassificationPROTEINS
Accuracy76.8
742
Node ClassificationChameleon
Accuracy43.6
549
Node ClassificationSquirrel
Accuracy42.2
500
Graph ClassificationNCI1
Accuracy85
460
Graph ClassificationIMDB-B
Accuracy80.2
322
Graph ClassificationENZYMES
Accuracy77.7
305
Node ClassificationCiteseer
Accuracy64.8
275
Graph ClassificationNCI109
Accuracy84.2
223
Graph ClassificationIMDB-M
Accuracy47.6
218
Graph RegressionZINC (test)
MAE0.024
204
Showing 10 of 74 rows
...

Other info

Code

Follow for update