Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Specformer: Spectral Graph Neural Networks Meet Transformers

About

Spectral graph neural networks (GNNs) learn graph representations via spectral-domain graph convolutions. However, most existing spectral graph filters are scalar-to-scalar functions, i.e., mapping a single eigenvalue to a single filtered value, thus ignoring the global pattern of the spectrum. Furthermore, these filters are often constructed based on some fixed-order polynomials, which have limited expressiveness and flexibility. To tackle these issues, we introduce Specformer, which effectively encodes the set of all eigenvalues and performs self-attention in the spectral domain, leading to a learnable set-to-set spectral filter. We also design a decoder with learnable bases to enable non-local graph convolution. Importantly, Specformer is equivariant to permutation. By stacking multiple Specformer layers, one can build a powerful spectral GNN. On synthetic datasets, we show that our Specformer can better recover ground-truth spectral filters than other spectral GNNs. Extensive experiments of both node-level and graph-level tasks on real-world graph datasets show that our Specformer outperforms state-of-the-art GNNs and learns meaningful spectrum patterns. Code and data are available at https://github.com/bdy9527/Specformer.

Deyu Bo, Chuan Shi, Lele Wang, Renjie Liao• 2023

Related benchmarks

TaskDatasetResultRank
Node ClassificationCora (test)
Mean Accuracy88.41
861
Node ClassificationChameleon
Accuracy49.79
640
Node ClassificationSquirrel
Accuracy38.24
591
Node ClassificationPubMed (test)
Accuracy89.19
546
Node ClassificationActor
Accuracy34.12
397
Node ClassificationChameleon (test)
Mean Accuracy73.31
297
Node ClassificationSquirrel (test)
Mean Accuracy62.15
267
Node ClassificationActor (test)
Mean Accuracy0.3826
237
Graph Classificationogbg-molpcba (test)
AP29.72
206
Graph RegressionZINC (test)
MAE0.066
204
Showing 10 of 25 rows

Other info

Follow for update