Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Graph Neural Networks Need Cluster-Normalize-Activate Modules

About

Graph Neural Networks (GNNs) are non-Euclidean deep learning models for graph-structured data. Despite their successful and diverse applications, oversmoothing prohibits deep architectures due to node features converging to a single fixed point. This severely limits their potential to solve complex tasks. To counteract this tendency, we propose a plug-and-play module consisting of three steps: Cluster-Normalize-Activate (CNA). By applying CNA modules, GNNs search and form super nodes in each layer, which are normalized and activated individually. We demonstrate in node classification and property prediction tasks that CNA significantly improves the accuracy over the state-of-the-art. Particularly, CNA reaches 94.18% and 95.75% accuracy on Cora and CiteSeer, respectively. It further benefits GNNs in regression tasks as well, reducing the mean squared error compared to all baselines. At the same time, GNNs with CNA require substantially fewer learnable parameters than competing architectures.

Arseny Skryagin, Felix Divo, Mohammad Amin Ali, Devendra Singh Dhami, Kristian Kersting• 2024

Related benchmarks

TaskDatasetResultRank
Node ClassificationCora
Accuracy94.18
1215
Node ClassificationPubmed
Accuracy90.16
819
Node ClassificationChameleon
Accuracy85.86
640
Node ClassificationWisconsin
Accuracy89.29
627
Node ClassificationTexas
Accuracy0.9
616
Node ClassificationSquirrel
Accuracy77.47
591
Node Classificationogbn-arxiv (test)
Accuracy74.64
433
Node ClassificationCiteseer
Accuracy95.75
393
Node ClassificationPhoto
Mean Accuracy95.96
343
Node ClassificationComputers
Mean Accuracy92.68
169
Showing 10 of 14 rows

Other info

Code

Follow for update