Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MedCoT: Medical Chain of Thought via Hierarchical Expert

About

Artificial intelligence has advanced in Medical Visual Question Answering (Med-VQA), but prevalent research tends to focus on the accuracy of the answers, often overlooking the reasoning paths and interpretability, which are crucial in clinical settings. Besides, current Med-VQA algorithms, typically reliant on singular models, lack the robustness needed for real-world medical diagnostics which usually require collaborative expert evaluation. To address these shortcomings, this paper presents MedCoT, a novel hierarchical expert verification reasoning chain method designed to enhance interpretability and accuracy in biomedical imaging inquiries. MedCoT is predicated on two principles: The necessity for explicit reasoning paths in Med-VQA and the requirement for multi-expert review to formulate accurate conclusions. The methodology involves an Initial Specialist proposing diagnostic rationales, followed by a Follow-up Specialist who validates these rationales, and finally, a consensus is reached through a vote among a sparse Mixture of Experts within the locally deployed Diagnostic Specialist, which then provides the definitive diagnosis. Experimental evaluations on four standard Med-VQA datasets demonstrate that MedCoT surpasses existing state-of-the-art approaches, providing significant improvements in performance and interpretability.

Jiaxiang Liu, Yuan Wang, Jiawei Du, Joey Tianyi Zhou, Zuozhu Liu• 2024

Related benchmarks

TaskDatasetResultRank
Medical Visual Question AnsweringSLAKE (test)
Overall Accuracy83.9
56
Medical Visual Question AnsweringPathVQA (test)
Accuracy66.2
55
Medical Visual Question AnsweringVQA-RAD (test)
Accuracy75.8
38
Medical Visual Question AnsweringPMC-VQA (test)
Accuracy80.3
36
Medical Visual Question AnsweringAggregate (SLAKE, VQA-RAD, PathVQA, PMC-VQA) Average (test)
Accuracy76.6
11
Showing 5 of 5 rows

Other info

Follow for update