Knowledge Graph Augmented Large Language Models for Disease Prediction
About
Electronic health records (EHRs) enable strong clinical prediction, but explanations are often coarse and hard to use for patient-level decisions. We propose a knowledge graph (KG)-guided chain-of-thought (CoT) framework for visit-level disease prediction on MIMIC-III. We map ICD-9 codes to PrimeKG, mine disease-relevant nodes and paths, and use these paths to scaffold temporally consistent CoT rationales, retaining only samples whose conclusions match observed outcomes. We fine-tune lightweight instruction-tuned LLMs (LLaMA-3.1-Instruct-8B and Gemma-7B) on two small cohorts (400 and 1,000 index visits) across ten PrimeKG-mapped diseases. Our models outperform strong classical baselines, reaching AUROC 0.66-0.70 and macro-AUPR 0.40-0.47. Without additional training, the models transfer zero-shot to the CRADLE cohort, improving accuracy from 0.40-0.51 to 0.72-0.77. In a blinded clinician study, KG-guided CoT rationales are consistently preferred for clarity, relevance, and correctness. Code is available at: https://github.com/JonathanWry/KG-guided-LLM-pipeline
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Clinical prediction | CRADLE | Accuracy76.86 | 17 | |
| visit-level disease prediction (10 diseases) | MIMIC III (400 index visits) | Accuracy83.89 | 11 | |
| visit-level disease prediction (10 diseases) | MIMIC III (1000 index visits) | Accuracy0.854 | 11 |