InstructUIE: Multi-task Instruction Tuning for Unified Information Extraction
About
Large language models have unlocked strong multi-task capabilities from reading instructive prompts. However, recent studies have shown that existing large models still have difficulty with information extraction tasks. For example, gpt-3.5-turbo achieved an F1 score of 18.22 on the Ontonotes dataset, which is significantly lower than the state-of-the-art performance. In this paper, we propose InstructUIE, a unified information extraction framework based on instruction tuning, which can uniformly model various information extraction tasks and capture the inter-task dependency. To validate the proposed method, we introduce IE INSTRUCTIONS, a benchmark of 32 diverse information extraction datasets in a unified text-to-text format with expert-written instructions. Experimental results demonstrate that our method achieves comparable performance to Bert in supervised settings and significantly outperforms the state-of-the-art and gpt3.5 in zero-shot settings.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Named Entity Recognition | CoNLL 03 | F1 (Entity)94.6 | 102 | |
| Named Entity Recognition | OntoNotes | F1-score88.64 | 102 | |
| Named Entity Recognition | Conll 2003 | F1 Score91.53 | 86 | |
| Named Entity Recognition | OntoNotes 5.0 | F1 Score90.19 | 79 | |
| Named Entity Recognition | BC5CDR | F1 Score91.9 | 70 | |
| Named Entity Recognition | MIT Movie | Entity F189.01 | 57 | |
| Relation Extraction | CONLL04 | Relation Strict F178.48 | 52 | |
| Named Entity Recognition | MIT Restaurant | -- | 50 | |
| Named Entity Recognition | tweetNER7 | Entity F166 | 49 | |
| Relation Extraction | SciERC | Relation Strict F145.15 | 48 |