Colorful Talks with Graphs: Human-Interpretable Graph Encodings for Large Language Models
About
Graph problems are fundamentally challenging for large language models (LLMs). While LLMs excel at processing unstructured text, graph tasks require reasoning over explicit structure, permutation invariance, and computationally complex relationships, creating a mismatch with the representations of text-based models. Our work investigates how LLMs can be effectively applied to graph problems despite these barriers. We introduce a human-interpretable structural encoding strategy for graph-to-text translation that injects graph structure directly into natural language prompts. Our method involves computing a variant of Weisfeiler-Lehman (WL) similarity classes and maps them to human-like color tokens rather than numeric labels. The key insight is that semantically meaningful and human-interpretable cues may be more effectively processed by LLMs than opaque symbolic encoding. Experimental results on multiple algorithmic and predictive graph tasks show the considerable improvements by our method on both synthetic and real-world datasets. By capturing both local and global-range dependencies, our method enhances LLM performance especially on graph tasks that require reasoning over global graph structure.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Cycle Check | BA and ER averaged (test) | Accuracy93 | 5 | |
| Shortest Path | BA and ER averaged (test) | Accuracy91.87 | 5 | |
| Maximum Flow | BA and ER averaged (test) | Accuracy36.59 | 5 | |
| Triangle Counting | BA and ER averaged (test) | Accuracy14.5 | 5 | |
| Node Classification | Cora 150 subgraphs total (30 sampled) | Macro F10.223 | 4 | |
| Node Classification | Citeseer 150 subgraphs total (30 sampled) | F1-Macro20.08 | 4 | |
| Node Classification | PubMed 150 subgraphs total (30 sampled) | F1-Macro14.33 | 4 | |
| Node Classification | OGBN-ArXiv 150 subgraphs total (30 sampled) | F1-Macro40.94 | 4 |