Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

BiomedGPT: A Generalist Vision-Language Foundation Model for Diverse Biomedical Tasks

About

Traditional biomedical artificial intelligence (AI) models, designed for specific tasks or modalities, often exhibit limited flexibility in real-world deployment and struggle to utilize holistic information. Generalist AI holds the potential to address these limitations due to its versatility in interpreting different data types and generating tailored outputs for diverse needs. However, existing biomedical generalist AI solutions are typically heavyweight and closed source to researchers, practitioners, and patients. Here, we propose BiomedGPT, the first open-source and lightweight vision-language foundation model, designed as a generalist capable of performing various biomedical tasks. BiomedGPT achieved state-of-the-art results in 16 out of 25 experiments while maintaining a computing-friendly model scale. We also conducted human evaluations to assess the capabilities of BiomedGPT in radiology visual question answering, report generation, and summarization. BiomedGPT exhibits robust prediction ability with a low error rate of 3.8% in question answering, satisfactory performance with an error rate of 8.3% in writing complex radiology reports, and competitive summarization ability with a nearly equivalent preference score to human experts. Our method demonstrates that effective training with diverse data can lead to more practical biomedical AI for improving diagnosis and workflow efficiency.

Kai Zhang, Rong Zhou, Eashan Adhikarla, Zhiling Yan, Yixin Liu, Jun Yu, Zhengliang Liu, Xun Chen, Brian D. Davison, Hui Ren, Jing Huang, Chen Chen, Yuyin Zhou, Sunyang Fu, Wei Liu, Tianming Liu, Xiang Li, Yong Chen, Lifang He, James Zou, Quanzheng Li, Hongfang Liu, Lichao Sun• 2023

Related benchmarks

TaskDatasetResultRank
Molecule CaptioningChEBI-20 (test)
BLEU-40.141
107
Radiology Report GenerationIU-Xray (test)
ROUGE-L0.285
55
Medical Report GenerationMIMIC-CXR (test)
ROUGE-L0.238
39
Medical Visual Question AnsweringSLAKE (test)
Closed Accuracy24.8
29
Medical Report GenerationMIMIC-CXR
F1 Score28.6
22
Medical Visual Question AnsweringMedical VQA Suite (MMMU-Med, VQA-RAD, SLAKE, PathVQA, PMC-VQA, OmniMedVQA, MedXpertQA)
MMMU-Med Score24.9
18
Medical Report GenerationIU X-Ray
Precision36
11
View ClassificationMulti-vendor TTE dataset (downstream)
Accuracy92.6
8
Image-Text RetrievalEchoGround-MIMIC (test)
Recall@50.0235
7
Disease ClassificationEchoGround-MIMIC (test)
AUC77.1
7
Showing 10 of 13 rows

Other info

Follow for update