Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Template-free Prompt Tuning for Few-shot NER

About

Prompt-based methods have been successfully applied in sentence-level few-shot learning tasks, mostly owing to the sophisticated design of templates and label words. However, when applied to token-level labeling tasks such as NER, it would be time-consuming to enumerate the template queries over all potential entity spans. In this work, we propose a more elegant method to reformulate NER tasks as LM problems without any templates. Specifically, we discard the template construction process while maintaining the word prediction paradigm of pre-training models to predict a class-related pivot word (or label word) at the entity position. Meanwhile, we also explore principled ways to automatically search for appropriate label words that the pre-trained models can easily adapt to. While avoiding complicated template-based process, the proposed LM objective also reduces the gap between different objectives used in pre-training and fine-tuning, thus it can better benefit the few-shot performance. Experimental results demonstrate the effectiveness of the proposed method over bert-tagger and template-based method under few-shot setting. Moreover, the decoding speed of the proposed method is up to 1930.12 times faster than the template-based method.

Ruotian Ma, Xin Zhou, Tao Gui, Yiding Tan, Linyang Li, Qi Zhang, Xuanjing Huang• 2021

Related benchmarks

TaskDatasetResultRank
Named Entity RecognitionCoNLL 03--
102
Named Entity RecognitionWnut 2017--
79
Named Entity RecognitionGUM
Micro F113.35
36
Named Entity RecognitionCoNLL (test)--
28
Named Entity RecognitionOntoNotes Onto-B 5.0
Micro-F135.7
26
Named Entity RecognitionOntoNotes Onto-C 5.0
Micro F128.8
26
Named Entity RecognitionOntoNotes Onto-A 5.0
Micro F121.29
26
Named Entity Recognitioni2b2 2014
Micro F1 Score0.1351
26
Showing 8 of 8 rows

Other info

Follow for update