BiomedCoOp: Learning to Prompt for Biomedical Vision-Language Models
About
Recent advancements in vision-language models (VLMs), such as CLIP, have demonstrated substantial success in self-supervised representation learning for vision tasks. However, effectively adapting VLMs to downstream applications remains challenging, as their accuracy often depends on time-intensive and expertise-demanding prompt engineering, while full model fine-tuning is costly. This is particularly true for biomedical images, which, unlike natural images, typically suffer from limited annotated datasets, unintuitive image contrasts, and nuanced visual features. Recent prompt learning techniques, such as Context Optimization (CoOp) intend to tackle these issues, but still fall short in generalizability. Meanwhile, explorations in prompt learning for biomedical image analysis are still highly limited. In this work, we propose BiomedCoOp, a novel prompt learning framework that enables efficient adaptation of BiomedCLIP for accurate and highly generalizable few-shot biomedical image classification. Our approach achieves effective prompt context learning by leveraging semantic consistency with average prompt ensembles from Large Language Models (LLMs) and knowledge distillation with a statistics-based prompt selection strategy. We conducted comprehensive validation of our proposed framework on 11 medical datasets across 9 modalities and 10 organs against existing state-of-the-art methods, demonstrating significant improvements in both accuracy and generalizability. The code is publicly available at https://github.com/HealthX-Lab/BiomedCoOp.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Medical Image Classification | BUSI | -- | 88 | |
| Image Classification | DermaMNIST | Accuracy62.59 | 23 | |
| Medical Image Classification | OCTMNIST | Accuracy66.93 | 19 | |
| Image Classification | Biomedical Datasets Average (test) | Accuracy72.42 | 18 | |
| Biomedical Image Classification | 11 Biomedical Datasets Average (test) | Avg Acc (K=1) (Biomed)57.03 | 10 | |
| Image Classification | Kvasir | Mean Accuracy78.89 | 7 | |
| Image Classification | Retina | Mean Accuracy61.28 | 7 | |
| Image Classification | CHMNIST | Mean Accuracy79.05 | 7 | |
| Image Classification | COVID-QU-Ex | Mean Accuracy78.72 | 7 | |
| Image Classification | LC25000 | Mean Accuracy0.9268 | 7 |