Graph Normalizing Flows
About
We introduce graph normalizing flows: a new, reversible graph neural network model for prediction and generation. On supervised tasks, graph normalizing flows perform similarly to message passing neural networks, but at a significantly reduced memory footprint, allowing them to scale to larger graphs. In the unsupervised case, we combine graph normalizing flows with a novel graph auto-encoder to create a generative model of graph structures. Our model is permutation-invariant, generating entire graphs with a single feed-forward pass, and achieves competitive results with the state-of-the art auto-regressive models, while being better suited to parallel computing architectures.
Jenny Liu, Aviral Kumar, Jimmy Ba, Jamie Kiros, Kevin Swersky• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Abstract graph generation | Community small | Degree0.2 | 17 | |
| Abstract graph generation | Ego small | Average MMD0.0437 | 17 | |
| Graph generation | Ego-small (test) | Degree0.03 | 11 | |
| Graph generation | Community small | MMD (Degree)0.2 | 8 |
Showing 4 of 4 rows