Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Scalable Deep Generative Modeling for Sparse Graphs

About

Learning graph generative models is a challenging task for deep learning and has wide applicability to a range of domains like chemistry, biology and social science. However current deep neural methods suffer from limited scalability: for a graph with $n$ nodes and $m$ edges, existing deep neural methods require $\Omega(n^2)$ complexity by building up the adjacency matrix. On the other hand, many real world graphs are actually sparse in the sense that $m\ll n^2$. Based on this, we develop a novel autoregressive model, named BiGG, that utilizes this sparsity to avoid generating the full adjacency matrix, and importantly reduces the graph generation time complexity to $O((n + m)\log n)$. Furthermore, during training this autoregressive model can be parallelized with $O(\log n)$ synchronization stages, which makes it much more efficient than other autoregressive models that require $\Omega(n)$. Experiments on several benchmarks show that the proposed approach not only scales to orders of magnitude larger graphs than previously possible with deep autoregressive graph generative models, but also yields better graph generation quality.

Hanjun Dai, Azade Nazi, Yujia Li, Bo Dai, Dale Schuurmans• 2020

Related benchmarks

TaskDatasetResultRank
Graph Generative ModelingPTC (test)
Degree Distribution Error1.00e-4
12
Graph generationTriangle Grid
MMD RBF0.12
12
Graph Generative ModelingMutag (test)
Degree Distribution0.004
12
Graph generationMUTAG
Training Time (s)5.2
10
Graph generationPTC
Train Time (s)7.66
10
Graph generationogbg-molbbbp
MMD RBF0.32
6
Graph generationPTC
MMD RBF0.04
6
Graph generationLobster
MMD RBF0.35
6
Graph generationPROTEIN
MMD RBF0.15
6
Graph generationProtein (test)
Training Time (s)130.3
6
Showing 10 of 14 rows

Other info

Follow for update