GMAI-VL & GMAI-VL-5.5M: A Large Vision-Language Model and A Comprehensive Multimodal Dataset Towards General Medical AI
About
Despite significant advancements in general AI, its effectiveness in the medical domain is limited by the lack of specialized medical knowledge. To address this, we formulate GMAI-VL-5.5M, a multimodal medical dataset created by converting hundreds of specialized medical datasets with various annotations into high-quality image-text pairs. This dataset offers comprehensive task coverage, diverse modalities, and rich image-text data. Building upon this dataset, we develop GMAI-VL, a general medical vision-language model, with a three-stage training strategy that enhances the integration of visual and textual information. This approach significantly improves the model's ability to process multimodal data, supporting accurate diagnoses and clinical decision-making. Experiments show that GMAI-VL achieves state-of-the-art performance across various multimodal medical tasks, including visual question answering and medical image diagnosis.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Medical Visual Question Answering | SLAKE (test) | -- | 29 | |
| Medical Visual Question Answering | VQA-RAD (test) | Accuracy64.6 | 13 | |
| Medical Visual Question Answering | PMC-VQA (test) | Accuracy52.3 | 13 | |
| Medical Visual Question Answering | PathVQA (test) | Accuracy47.2 | 13 | |
| Medical Visual Question Answering | MMMU Health & Medicine (test) | Accuracy51.2 | 12 |