Representation Learning on Graphs with Jumping Knowledge Networks
About
Recent deep learning approaches for representation learning on graphs follow a neighborhood aggregation procedure. We analyze some important properties of these models, and propose a strategy to overcome those. In particular, the range of "neighboring" nodes that a node's representation draws from strongly depends on the graph structure, analogous to the spread of a random walk. To adapt to local neighborhood properties and tasks, we explore an architecture -- jumping knowledge (JK) networks -- that flexibly leverages, for each node, different neighborhood ranges to enable better structure-aware representation. In a number of experiments on social, bioinformatics and citation networks, we demonstrate that our model achieves state-of-the-art performance. Furthermore, combining the JK framework with models like Graph Convolutional Networks, GraphSAGE and Graph Attention Networks consistently improves those models' performance.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Node Classification | Cora | Accuracy87.46 | 885 | |
| Node Classification | Citeseer | Accuracy75.99 | 804 | |
| Node Classification | Pubmed | Accuracy89.45 | 742 | |
| Node Classification | Citeseer (test) | Accuracy0.783 | 729 | |
| Node Classification | Cora (test) | Mean Accuracy89.1 | 687 | |
| Node Classification | Chameleon | Accuracy63.79 | 549 | |
| Node Classification | PubMed (test) | Accuracy85.8 | 500 | |
| Node Classification | Squirrel | Accuracy45.03 | 500 | |
| Node Classification | Cornell | Accuracy75.68 | 426 | |
| Node Classification | Texas | Accuracy62.7 | 410 |