Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Label Verbalization and Entailment for Effective Zero- and Few-Shot Relation Extraction

About

Relation extraction systems require large amounts of labeled examples which are costly to annotate. In this work we reformulate relation extraction as an entailment task, with simple, hand-made, verbalizations of relations produced in less than 15 min per relation. The system relies on a pretrained textual entailment engine which is run as-is (no training examples, zero-shot) or further fine-tuned on labeled examples (few-shot or fully trained). In our experiments on TACRED we attain 63% F1 zero-shot, 69% with 16 examples per relation (17% points better than the best supervised system on the same conditions), and only 4 points short to the state-of-the-art (which uses 20 times more training data). We also show that the performance can be improved significantly with larger entailment models, up to 12 points in zero-shot, allowing to report the best results to date on TACRED when fully trained. The analysis shows that our few-shot systems are specially effective when discriminating between relations, and that the performance difference in low data regimes comes mainly from identifying no-relation cases.

Oscar Sainz, Oier Lopez de Lacalle, Gorka Labaka, Ander Barrena, Eneko Agirre• 2021

Related benchmarks

TaskDatasetResultRank
Relation ExtractionTACRED (test)
F1 Score73.9
194
Natural Language InferenceMNLI (matched)
Accuracy91.7
110
Relation ExtractionSemEval (test)
Micro F123.7
55
Relation ExtractionTACREV (test)
F1 Score57.2
27
Relation ExtractionTACRED v1.0 (5% train)
Micro F10.69
19
Relation ExtractionRETACRED (test)
Precision71.7
17
Relation ExtractionTACRED v1.0 (full)
Micro F173.9
16
Relation ClassificationTACRED
F1 Score73.9
14
Relation ExtractionTACRED 1% v1.0 (train)
Micro F163.7
13
Relation ExtractionTACRED v1.0 (10% train)
Micro F167.9
13
Showing 10 of 20 rows

Other info

Code

Follow for update