Multimodal Cancer Survival Analysis via Hypergraph Learning with Cross-Modality Rebalance
About
Multimodal pathology-genomic analysis has become increasingly prominent in cancer survival prediction. However, existing studies mainly utilize multi-instance learning to aggregate patch-level features, neglecting the information loss of contextual and hierarchical details within pathology images. Furthermore, the disparity in data granularity and dimensionality between pathology and genomics leads to a significant modality imbalance. The high spatial resolution inherent in pathology data renders it a dominant role while overshadowing genomics in multimodal integration. In this paper, we propose a multimodal survival prediction framework that incorporates hypergraph learning to effectively capture both contextual and hierarchical details from pathology images. Moreover, it employs a modality rebalance mechanism and an interactive alignment fusion strategy to dynamically reweight the contributions of the two modalities, thereby mitigating the pathology-genomics imbalance. Quantitative and qualitative experiments are conducted on five TCGA datasets, demonstrating that our model outperforms advanced methods by over 3.4\% in C-Index performance.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Glioma Grading | TCGA GBM-LGG (3-fold val) | AUC88.7 | 48 | |
| Survival Prediction | TCGA GBM-LGG Internal (test) | C-Index76.42 | 37 | |
| Survival Prediction | CPTAC External (test) | C-Index53.76 | 27 | |
| Diagnosis | TCGA GBM-LGG and IvyGAP (3-fold val) | AUC95.11 | 26 |