Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Contrastive Multi-View Representation Learning on Graphs

About

We introduce a self-supervised approach for learning node and graph level representations by contrasting structural views of graphs. We show that unlike visual representation learning, increasing the number of views to more than two or contrasting multi-scale encodings do not improve performance, and the best performance is achieved by contrasting encodings from first-order neighbors and a graph diffusion. We achieve new state-of-the-art results in self-supervised learning on 8 out of 8 node and graph classification benchmarks under the linear evaluation protocol. For example, on Cora (node) and Reddit-Binary (graph) classification benchmarks, we achieve 86.8% and 84.5% accuracy, which are 5.5% and 2.4% relative improvements over previous state-of-the-art. When compared to supervised baselines, our approach outperforms them in 4 out of 8 benchmarks. Source code is released at: https://github.com/kavehhassani/mvgrl

Kaveh Hassani, Amir Hosein Khasahmadi• 2020

Related benchmarks

TaskDatasetResultRank
Node ClassificationCora
Accuracy86.8
1215
Graph ClassificationPROTEINS
Accuracy74.2
994
Node ClassificationCiteseer
Accuracy75.02
931
Graph ClassificationMUTAG
Accuracy89.7
862
Node ClassificationCora (test)
Mean Accuracy87.36
861
Node ClassificationCiteseer (test)
Accuracy0.787
824
Node ClassificationPubmed
Accuracy87.12
819
Node ClassificationChameleon
Accuracy54.61
640
Node ClassificationTexas
Accuracy0.697
616
Node ClassificationSquirrel
Accuracy39.9
591
Showing 10 of 114 rows
...

Other info

Follow for update