HealthGPT: A Medical Large Vision-Language Model for Unifying Comprehension and Generation via Heterogeneous Knowledge Adaptation
About
We present HealthGPT, a powerful Medical Large Vision-Language Model (Med-LVLM) that integrates medical visual comprehension and generation capabilities within a unified autoregressive paradigm. Our bootstrapping philosophy is to progressively adapt heterogeneous comprehension and generation knowledge to pre-trained large language models (LLMs). This is achieved through a novel heterogeneous low-rank adaptation (H-LoRA) technique, which is complemented by a tailored hierarchical visual perception approach and a three-stage learning strategy. To effectively learn the HealthGPT, we devise a comprehensive medical domain-specific comprehension and generation dataset called VL-Health. Experimental results demonstrate exceptional performance and scalability of HealthGPT in medical visual unified tasks. Our project can be accessed at https://github.com/DCDmllm/HealthGPT.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Medical Visual Question Answering | VQA-RAD | -- | 198 | |
| Medical Visual Question Answering | PathVQA | -- | 86 | |
| Medical Visual Question Answering | OmniMedVQA (test) | CT Accuracy70.3 | 50 | |
| Open-ended VQA | MMOral-OPG | Teeth Accuracy30.64 | 38 | |
| Multimodal Dental Image Analysis | MMOral-Uni 1.0 (test) | Loc Score18 | 28 | |
| Open-ended VQA | MMOral-X | Simple Score6.34 | 21 | |
| Medical Visual Question Answering | Slake | Closed Accuracy71.9 | 17 | |
| Tumor analysis | TumorCoT 1.5M (test) | Organ Position30.58 | 17 | |
| Fundus reading | FunBench | Accuracy52.4 | 14 | |
| Fundus reading | GMAI-Fundus | Accuracy46.3 | 14 |