Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

InstructUIE: Multi-task Instruction Tuning for Unified Information Extraction

About

Large language models have unlocked strong multi-task capabilities from reading instructive prompts. However, recent studies have shown that existing large models still have difficulty with information extraction tasks. For example, gpt-3.5-turbo achieved an F1 score of 18.22 on the Ontonotes dataset, which is significantly lower than the state-of-the-art performance. In this paper, we propose InstructUIE, a unified information extraction framework based on instruction tuning, which can uniformly model various information extraction tasks and capture the inter-task dependency. To validate the proposed method, we introduce IE INSTRUCTIONS, a benchmark of 32 diverse information extraction datasets in a unified text-to-text format with expert-written instructions. Experimental results demonstrate that our method achieves comparable performance to Bert in supervised settings and significantly outperforms the state-of-the-art and gpt3.5 in zero-shot settings.

Xiao Wang, Weikang Zhou, Can Zu, Han Xia, Tianze Chen, Yuansen Zhang, Rui Zheng, Junjie Ye, Qi Zhang, Tao Gui, Jihua Kang, Jingsheng Yang, Siyuan Li, Chunsai Du• 2023

Related benchmarks

TaskDatasetResultRank
Named Entity RecognitionCoNLL 03
F1 (Entity)94.6
102
Named Entity RecognitionOntoNotes
F1-score88.64
102
Named Entity RecognitionConll 2003
F1 Score91.53
86
Named Entity RecognitionOntoNotes 5.0
F1 Score90.19
79
Named Entity RecognitionBC5CDR
F1 Score91.9
70
Named Entity RecognitionMIT Movie
Entity F189.01
57
Relation ExtractionCONLL04
Relation Strict F178.48
52
Named Entity RecognitionMIT Restaurant--
50
Named Entity RecognitiontweetNER7
Entity F166
49
Relation ExtractionSciERC
Relation Strict F145.15
48
Showing 10 of 74 rows
...

Other info

Code

Follow for update