Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Vision-Language Foundation Model to Enhance Efficiency of Chest X-ray Interpretation

About

Over 1.4 billion chest X-rays (CXRs) are performed annually due to their cost-effectiveness as an initial diagnostic test. This scale of radiological studies provides a significant opportunity to streamline CXR interpretation and documentation. While foundation models are a promising solution, the lack of publicly available large-scale datasets and benchmarks inhibits their iterative development and real-world evaluation. To overcome these challenges, we constructed a large-scale dataset (CheXinstruct), which we utilized to train a vision-language foundation model (CheXagent). We systematically demonstrated competitive performance across eight distinct task types on our novel evaluation benchmark (CheXbench). Beyond technical validation, we assessed the real-world utility of CheXagent in directly drafting radiology reports. Our clinical assessment with eight radiologists revealed a 36% time saving for residents using CheXagent-drafted reports, while attending radiologists showed no significant time difference editing resident-drafted or CheXagent-drafted reports. The CheXagent-drafted reports improved the writing efficiency of both radiology residents and attending radiologists in 81% and 61% of cases, respectively, without loss of quality. Overall, we demonstrate that CheXagent can effectively perform a variety of CXR interpretation tasks and holds potential to assist radiologists in routine clinical workflows.

Zhihong Chen, Maya Varma, Justin Xu, Magdalini Paschali, Dave Van Veen, Andrew Johnston, Alaa Youssef, Louis Blankemeier, Christian Bluethgen, Stephan Altmayer, Jeya Maria Jose Valanarasu, Mohamed Siddig Eltayeb Muneer, Eduardo Pontes Reis, Joseph Paul Cohen, Cameron Olsen, Tanishq Mathew Abraham, Emily B. Tsai, Christopher F. Beaulieu, Jenia Jitsev, Sergios Gatidis, Jean-Benoit Delbrouck, Akshay S. Chaudhari, Curtis P. Langlotz• 2024

Related benchmarks

TaskDatasetResultRank
Radiology Report GenerationMIMIC-CXR (test)
BLEU-40.047
121
Image ClassificationCovidx
Accuracy34.3
57
Visual Question AnsweringChest X-ray VQA (test)
Overall Accuracy47.41
43
Medical Report GenerationMIMIC-CXR
F1 Score31.95
22
Chest X-ray Report GenerationMIMIC-CXR (test)
F1 Macro (14)38.9
21
Medical Image Report LabelingMIMIC-CXR (test)
Macro F1 (14 Labels)38.9
21
Radiology Report GenerationRadVLM MIMIC-CXR (test)
ROUGE-L22.5
13
Medical Report GenerationIU X-Ray
Precision50.37
11
Abnormality DetectionCXR
IoU31
8
Close-Ended Visual Question AnsweringCXR
BERTScore90
8
Showing 10 of 23 rows

Other info

Follow for update