Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

A Continued Pretrained LLM Approach for Automatic Medical Note Generation

About

LLMs are revolutionizing NLP tasks. However, the use of the most advanced LLMs, such as GPT-4, is often prohibitively expensive for most specialized fields. We introduce HEAL, the first continuously trained 13B LLaMA2-based LLM that is purpose-built for medical conversations and measured on automated scribing. Our results demonstrate that HEAL outperforms GPT-4 and PMC-LLaMA in PubMedQA, with an accuracy of 78.4\%. It also achieves parity with GPT-4 in generating medical notes. Remarkably, HEAL surpasses GPT-4 and Med-PaLM 2 in identifying more correct medical concepts and exceeds the performance of human scribes and other comparable models in correctness and completeness.

Dong Yuan, Eti Rastogi, Gautam Naik, Sree Prasanna Rajagopal, Sagar Goyal, Fen Zhao, Bharath Chintagunta, Jeff Ward• 2024

Related benchmarks

TaskDatasetResultRank
Medical Data and Knowledge ProcessingEHRStruct eICU
D-U1 Accuracy11
20
Data-Driven Structured EHR Understanding and ReasoningSynthea
D-R2 Accuracy4
19
Showing 2 of 2 rows

Other info

Follow for update