Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Masks and Manuscripts: Advancing Medical Pre-training with End-to-End Masking and Narrative Structuring

About

Contemporary medical contrastive learning faces challenges from inconsistent semantics and sample pair morphology, leading to dispersed and converging semantic shifts. The variability in text reports, due to multiple authors, complicates semantic consistency. To tackle these issues, we propose a two-step approach. Initially, text reports are converted into a standardized triplet format, laying the groundwork for our novel concept of ``observations'' and ``verdicts''. This approach refines the {Entity, Position, Exist} triplet into binary questions, guiding towards a clear ``verdict''. We also innovate in visual pre-training with a Meijering-based masking, focusing on features representative of medical images' local context. By integrating this with our text conversion method, our model advances cross-modal representation in a multimodal contrastive learning framework, setting new benchmarks in medical image analysis.

Shreyank N Gowda, David A. Clifton• 2024

Related benchmarks

TaskDatasetResultRank
Medical Image SegmentationRSNA Pneumonia
Dice Score76.68
49
ClassificationCheXpert (test)
AUC ROC90.88
48
Image ClassificationSIIM-ACR (test)
AUROC93.88
45
ClassificationRSNA Pneumonia (test)
AUC-ROC0.9191
27
SegmentationSIIM-ACR
Dice Score80.28
27
ClassificationRSNA Pneumonia
Accuracy83.14
21
Image ClassificationNIH ChestX-ray
Accuracy88.52
21
Medical Image SegmentationCOVID-19
Dice Score45.04
21
Image ClassificationSIIM-ACR
Accuracy86.15
12
ClassificationCovid-19 CXR
AUC75.15
5
Showing 10 of 11 rows

Other info

Follow for update