Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

NuNER: Entity Recognition Encoder Pre-training via LLM-Annotated Data

About

Large Language Models (LLMs) have shown impressive abilities in data annotation, opening the way for new approaches to solve classic NLP problems. In this paper, we show how to use LLMs to create NuNER, a compact language representation model specialized in the Named Entity Recognition (NER) task. NuNER can be fine-tuned to solve downstream NER problems in a data-efficient way, outperforming similar-sized foundation models in the few-shot regime and competing with much larger LLMs. We find that the size and entity-type diversity of the pre-training dataset are key to achieving good performance. We view NuNER as a member of the broader family of task-specific foundation models, recently unlocked by LLMs.

Sergei Bogdanov, Alexandre Constantin, Timoth\'ee Bernard, Benoit Crabb\'e, Etienne Bernard• 2024

Related benchmarks

TaskDatasetResultRank
Named Entity RecognitionCoNLL 03--
102
Named Entity RecognitionFew-NERD INTER 1.0 (test)
Average F169.61
62
Named Entity RecognitionMIT Restaurant
Micro-F168.57
50
Named Entity RecognitionFewNERD INTRA--
47
Extractive Question AnsweringSQuAD 2.0
F1 Score52.67
34
Entity recognitionMIT Restaurant
F1-macro84.79
24
Entity recognitionBioNLP
F1 Macro80.85
24
Relation ExtractionCoNLL 04
F165.12
24
Named Entity RecognitionMIT Movie
Entity F164.88
22
Entity recognitionBioNLP2004, MIT Movie, MIT Restaurant, and OntoNotes 5.0 (Average) few-shot (test)
Avg F1 (few-shot)71.53
21
Showing 10 of 17 rows

Other info

Code

Follow for update