Hierarchical Multi-Scale Graph Learning with Knowledge-Guided Attention for Whole-Slide Image Survival Analysis
About
We propose a Hierarchical Multi-scale Knowledge-aware Graph Network (HMKGN) that models multi-scale interactions and spatially hierarchical relationships within whole-slide images (WSIs) for cancer prognostication. Unlike conventional attention-based MIL, which ignores spatial organization, or graph-based MIL, which relies on static handcrafted graphs, HMKGN enforces a hierarchical structure with spatial locality constraints, wherein local cellular-level dynamic graphs aggregate spatially proximate patches within each region of interest (ROI) and a global slide-level dynamic graph integrates ROI-level features into WSI-level representations. Moreover, multi-scale integration at the ROI level combines coarse contextual features from broader views with fine-grained structural representations from local patch-graph aggregation. We evaluate HMKGN on four TCGA cohorts (KIRC, LGG, PAAD, and STAD; N=513, 487, 138, and 370) for survival prediction. It consistently outperforms existing MIL-based models, yielding improved concordance indices (10.85% better) and statistically significant stratification of patient survival risk (log-rank p < 0.05).
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Survival Prediction | TCGA-STAD | C-index0.6411 | 62 | |
| Survival Prediction | KIRC TCGA | C-Index0.7486 | 57 | |
| Survival Prediction | TCGA LGG | C-index0.7699 | 15 | |
| Survival Prediction | TCGA-PAAD | C-index0.6897 | 7 | |
| Survival Prediction | TCGA Mean of KIRC, LGG, PAAD, STAD | -- | 6 |