Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

M-FLAG: Medical Vision-Language Pre-training with Frozen Language Models and Latent Space Geometry Optimization

About

Medical vision-language models enable co-learning and integrating features from medical imaging and clinical text. However, these models are not easy to train and the latent representation space can be complex. Here we propose a novel way for pre-training and regularising medical vision-language models. The proposed method, named Medical vision-language pre-training with Frozen language models and Latent spAce Geometry optimization (M-FLAG), leverages a frozen language model for training stability and efficiency and introduces a novel orthogonality loss to harmonize the latent space geometry. We demonstrate the potential of the pre-trained model on three downstream tasks: medical image classification, segmentation, and object detection. Extensive experiments across five public datasets demonstrate that M-FLAG significantly outperforms existing medical vision-language pre-training approaches and reduces the number of parameters by 78\%. Notably, M-FLAG achieves outstanding performance on the segmentation task while using only 1\% of the RSNA dataset, even outperforming ImageNet pre-trained models that have been fine-tuned using 100\% of the data.

Che Liu, Sibo Cheng, Chen Chen, Mengyun Qiao, Weitong Zhang, Anand Shah, Wenjia Bai, Rossella Arcucci• 2023

Related benchmarks

TaskDatasetResultRank
Object DetectionRSNA
mAP (%)25.4
106
Semantic segmentationSIIM
Dice Coefficient (%)64.8
96
Semantic segmentationRSNA
Dice Score70.5
90
Object DetectionObject-CXR
mAP19.5
58
Chest X-ray classificationNIH (test)
AUROC84
47
ClassificationRSNA (test)
F1 Score64.4
44
Linear ClassificationRSNA (test)
AUC90.5
39
Linear ClassificationCOVIDx (test)
Accuracy90.7
39
Linear ClassificationCheXpert (test)
AUC0.886
39
Image ClassificationSIIM (test)
F1 Score72.1
30
Showing 10 of 23 rows

Other info

Code

Follow for update