Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Hyper-SAGNN: a self-attention based graph neural network for hypergraphs

About

Graph representation learning for hypergraphs can be used to extract patterns among higher-order interactions that are critically important in many real world problems. Current approaches designed for hypergraphs, however, are unable to handle different types of hypergraphs and are typically not generic for various learning tasks. Indeed, models that can predict variable-sized heterogeneous hyperedges have not been available. Here we develop a new self-attention based graph neural network called Hyper-SAGNN applicable to homogeneous and heterogeneous hypergraphs with variable hyperedge sizes. We perform extensive evaluations on multiple datasets, including four benchmark network datasets and two single-cell Hi-C datasets in genomics. We demonstrate that Hyper-SAGNN significantly outperforms the state-of-the-art methods on traditional tasks while also achieving great performance on a new task called outsider identification. Hyper-SAGNN will be useful for graph representation learning to uncover complex higher-order interactions in different applications.

Ruochi Zhang, Yuesong Zou, Jian Ma• 2019

Related benchmarks

TaskDatasetResultRank
Node ClassificationPrimary School
Mean Accuracy83.27
16
Node ClassificationAverage Performance High School, Primary School
Mean Accuracy0.839
16
Node ClassificationHigh School
Mean ACC84.89
16
Hyperedge predictionEmail Enron (Transductive)
Mean AUC83.59
9
Hyperedge predictionHigh School Transductive
Mean AUC94.98
9
Hyperedge predictionPrimary School (Transductive)
Mean AUC90.77
9
Hyperedge predictionCongress Bill (Transductive)
Mean AUC82.84
9
Hyperedge predictionEmail Eu (Transductive)
Mean AUC79.61
9
Hyperedge predictionNDC Class Transductive
Mean AUC80.76
9
Hyperedge predictionUsers-Threads (Transductive)
Mean AUC79.62
9
Showing 10 of 42 rows

Other info

Follow for update