Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

BioMegatron: Larger Biomedical Domain Language Model

About

There has been an influx of biomedical domain-specific language models, showing language models pre-trained on biomedical text perform better on biomedical domain benchmarks than those trained on general domain text corpora such as Wikipedia and Books. Yet, most works do not study the factors affecting each domain language application deeply. Additionally, the study of model size on domain-specific models has been mostly missing. We empirically study and evaluate several factors that can affect performance on domain language applications, such as the sub-word vocabulary set, model size, pre-training corpus, and domain transfer. We show consistent improvements on benchmarks with our larger BioMegatron model trained on a larger domain corpus, contributing to our understanding of domain language model applications. We demonstrate noticeable improvements over the previous state-of-the-art (SOTA) on standard biomedical NLP benchmarks of named entity recognition, relation extraction, and question answering. Model checkpoints and code are available at [https://ngc.nvidia.com] and [https://github.com/NVIDIA/NeMo].

Hoo-Chang Shin, Yang Zhang, Evelina Bakhturina, Raul Puri, Mostofa Patwary, Mohammad Shoeybi, Raghav Mani• 2020

Related benchmarks

TaskDatasetResultRank
Question AnsweringSQuAD v1.1 (dev)
F1 Score91.8
375
Question AnsweringSQuAD v2.0 (dev)
F186.4
158
Named Entity RecognitionBC5CDR (test)
Macro F1 (span-level)92.9
80
Named Entity RecognitionNCBI-disease
F1 Score87.8
29
Question AnsweringBioASQ factoid 7b (test)
SAcc47.4
13
Relation ExtractionChemProt
F1 Score77
10
Showing 6 of 6 rows

Other info

Code

Follow for update