Text-Free Multi-domain Graph Pre-training: Toward Graph Foundation Models
About
Given the ubiquity of graph data, it is intriguing to ask: Is it possible to train a graph foundation model on a broad range of graph data across diverse domains? A major hurdle toward this goal lies in the fact that graphs from different domains often exhibit profoundly divergent characteristics. Although there have been some initial efforts in integrating multi-domain graphs for pre-training, they primarily rely on textual descriptions to align the graphs, limiting their application to text-attributed graphs. Moreover, different source domains may conflict or interfere with each other, and their relevance to the target domain can vary significantly. To address these issues, we propose MDGPT, a text free Multi-Domain Graph Pre-Training and adaptation framework designed to exploit multi-domain knowledge for graph learning. First, we propose a set of domain tokens to to align features across source domains for synergistic pre-training. Second, we propose a dual prompts, consisting of a unifying prompt and a mixing prompt, to further adapt the target domain with unified multi-domain knowledge and a tailored mixture of domain-specific knowledge. Finally, we conduct extensive experiments involving six public datasets to evaluate and analyze MDGPT, which outperforms prior art by up to 37.9%.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Node Classification | Cora | Accuracy60.6 | 1215 | |
| Node Classification | Cora (test) | Mean Accuracy60 | 861 | |
| Node Classification | ogbn-arxiv (test) | Accuracy50.1 | 433 | |
| Node Classification | Pubmed | Accuracy58.7 | 396 | |
| Node Classification | Citeseer | Accuracy55.9 | 393 | |
| Node Classification | wikiCS | Accuracy54.1 | 317 | |
| Node Classification | Ogbn-arxiv | Accuracy53.4 | 170 | |
| Graph Classification | Pubmed | Accuracy67.6 | 101 | |
| Graph Classification | Citeseer | Accuracy59.3 | 99 | |
| Graph Classification | Wiki CS | Accuracy56 | 96 |