Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Evidence-based diagnostic reasoning with multi-agent copilot for human pathology

About

Pathology is experiencing rapid digital transformation driven by whole-slide imaging and artificial intelligence (AI). While deep learning-based computational pathology has achieved notable success, traditional models primarily focus on image analysis without integrating natural language instruction or rich, text-based context. Current multimodal large language models (MLLMs) in computational pathology face limitations, including insufficient training data, inadequate support and evaluation for multi-image understanding, and a lack of autonomous, diagnostic reasoning capabilities. To address these limitations, we introduce PathChat+, a new MLLM specifically designed for human pathology, trained on over 1 million diverse, pathology-specific instruction samples and nearly 5.5 million question answer turns. Extensive evaluations across diverse pathology benchmarks demonstrated that PathChat+ substantially outperforms the prior PathChat copilot, as well as both state-of-the-art (SOTA) general-purpose and other pathology-specific models. Furthermore, we present SlideSeek, a reasoning-enabled multi-agent AI system leveraging PathChat+ to autonomously evaluate gigapixel whole-slide images (WSIs) through iterative, hierarchical diagnostic reasoning, reaching high accuracy on DDxBench, a challenging open-ended differential diagnosis benchmark, while also capable of generating visually grounded, humanly-interpretable summary reports.

Luca L. Weishaupt, Chengkuan Chen, Drew F. K. Williamson, Richard J. Chen, Guillaume Jaume, Tong Ding, Bowen Chen, Anurag Vaidya, Long Phi Le, Guillaume Jaume, Ming Y. Lu, Faisal Mahmood• 2025

Related benchmarks

TaskDatasetResultRank
Multiple-choice Question AnsweringPathMMU Atlas tiny (test)
Accuracy82.7
13
Multiple-choice Question AnsweringPathMMU EduContent n=255 (test-tiny)
Accuracy76.9
13
Multiple-choice Visual Question AnsweringPathMMU SocialPath tiny n=229 (test)
Accuracy75.1
13
Visual Question AnsweringPathMMU All-tiny (test)
Accuracy77.1
13
Multiple-Choice QuestionsPathMMU PathCLS n = 177 (test-tiny)
Accuracy74
13
Multiple-choice Visual Question AnsweringPathMMU PubMed (test-tiny)
Accuracy76.9
13
Image CaptioningPathQABench Caption
METEOR28.1
12
Multiple-choice Question AnsweringPathMMU Atlas (test)
Accuracy84
12
Multiple-choice Question AnsweringPathMMU EduContent n=1,938 (test-all)
Accuracy73.5
12
Multiple-choice Question AnsweringBRACS n = 570 (test)
Accuracy63.3
12
Showing 10 of 19 rows

Other info

Follow for update