Semi-supervised Multitask Learning for Sequence Labeling
About
We propose a sequence labeling framework with a secondary training objective, learning to predict surrounding words for every word in the dataset. This language modeling objective incentivises the system to learn general-purpose patterns of semantic and syntactic composition, which are also useful for improving accuracy on different sequence labeling tasks. The architecture was evaluated on a range of datasets, covering the tasks of error detection in learner texts, named entity recognition, chunking and POS-tagging. The novel language modeling objective provided consistent performance improvements on every benchmark, without requiring any additional annotated or unannotated data.
Marek Rei• 2017
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Named Entity Recognition | CoNLL 2003 (test) | F1 Score87.94 | 539 | |
| Chunking | CoNLL 2000 (test) | F1 Score94.33 | 88 | |
| Named Entity Recognition | Conll 2003 | F1 Score86.26 | 86 | |
| Part-of-Speech Tagging | WSJ (test) | Accuracy97.14 | 51 | |
| Error detection | FCE (test) | F0.5 Score48.48 | 16 | |
| Error detection | CoNLL-14 (test1) | P17.68 | 9 | |
| Error detection | CoNLL-14 (test2) | Precision27.62 | 9 |
Showing 7 of 7 rows