Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Bidirectional LSTM-CRF Models for Sequence Tagging

About

In this paper, we propose a variety of Long Short-Term Memory (LSTM) based models for sequence tagging. These models include LSTM networks, bidirectional LSTM (BI-LSTM) networks, LSTM with a Conditional Random Field (CRF) layer (LSTM-CRF) and bidirectional LSTM with a CRF layer (BI-LSTM-CRF). Our work is the first to apply a bidirectional LSTM CRF (denoted as BI-LSTM-CRF) model to NLP benchmark sequence tagging data sets. We show that the BI-LSTM-CRF model can efficiently use both past and future input features thanks to a bidirectional LSTM component. It can also use sentence level tag information thanks to a CRF layer. The BI-LSTM-CRF model can produce state of the art (or close to) accuracy on POS, chunking and NER data sets. In addition, it is robust and has less dependence on word embedding as compared to previous observations.

Zhiheng Huang, Wei Xu, Kai Yu• 2015

Related benchmarks

TaskDatasetResultRank
Named Entity RecognitionCoNLL 2003 (test)
F1 Score90.1
539
Named Entity RecognitionCoNLL English 2003 (test)
F1 Score90.1
135
Named Entity RecognitionCoNLL Spanish NER 2002 (test)
F1 Score80.33
98
ChunkingCoNLL 2000 (test)
F1 Score94.46
88
Named Entity RecognitionCoNLL Dutch 2002 (test)
F1 Score79.87
87
Named Entity RecognitionConll 2003
F1 Score90.1
86
Named Entity RecognitionCoNLL German 2003 (test)
F1 Score73.42
78
Part-of-Speech TaggingPenn Treebank (test)
Accuracy97.55
64
Named Entity RecognitionMSRA (test)
F1 Score88.81
63
Named Entity RecognitionOntoNotes 4.0 (test)
F1 Score64.3
55
Showing 10 of 30 rows

Other info

Follow for update