Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Semi-supervised sequence tagging with bidirectional language models

About

Pre-trained word embeddings learned from unlabeled text have become a standard component of neural network architectures for NLP tasks. However, in most cases, the recurrent network that operates on word-level representations to produce context sensitive representations is trained on relatively little labeled data. In this paper, we demonstrate a general semi-supervised approach for adding pre- trained context embeddings from bidirectional language models to NLP systems and apply it to sequence labeling tasks. We evaluate our model on two standard datasets for named entity recognition (NER) and chunking, and in both cases achieve state of the art results, surpassing previous systems that use other forms of transfer or joint learning with additional labeled data and task specific gazetteers.

Matthew E. Peters, Waleed Ammar, Chandra Bhagavatula, Russell Power• 2017

Related benchmarks

TaskDatasetResultRank
Named Entity RecognitionCoNLL 2003 (test)
F1 Score92.2
539
Language ModelingOne Billion Word Benchmark (test)
Test Perplexity47.5
108
ChunkingCoNLL 2000 (test)
F1 Score96.37
88
Named Entity RecognitionConll 2003
F1 Score91.93
86
Named Entity RecognitionNER (test)
F1 Score91.93
68
ChunkingChunk (test)
F1 Score96.37
28
Entity recognitionSCIERC (test)
F1 Score62
20
Keyphrase ExtractionSemEval Task 10 ScienceIE 2017 (test)
F1 Score44
15
Entity recognitionSCIERC (dev)
Precision67.2
6
Span IdentificationSemEval ScienceIE Task 10 2017 (test)
F1 Score55
3
Showing 10 of 12 rows

Other info

Follow for update