Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ActiveAED: A Human in the Loop Improves Annotation Error Detection

About

Manually annotated datasets are crucial for training and evaluating Natural Language Processing models. However, recent work has discovered that even widely-used benchmark datasets contain a substantial number of erroneous annotations. This problem has been addressed with Annotation Error Detection (AED) models, which can flag such errors for human re-annotation. However, even though many of these AED methods assume a final curation step in which a human annotator decides whether the annotation is erroneous, they have been developed as static models without any human-in-the-loop component. In this work, we propose ActiveAED, an AED method that can detect errors more accurately by repeatedly querying a human for error corrections in its prediction loop. We evaluate ActiveAED on eight datasets spanning five different tasks and find that it leads to improvements over the state of the art on seven of them, with gains of up to six percentage points in average precision.

Leon Weber, Barbara Plank• 2023

Related benchmarks

TaskDatasetResultRank
LF Mislabeling IdentificationIMDB
AP36.6
38
Annotation Error DetectionATIS
AP98.7
6
Annotation Error DetectionSI-Flights
AP86.6
6
Annotation Error DetectionSST
AP53
6
Annotation Error DetectionConll 2003
AP33.3
6
Annotation Error DetectionSI-Companies
AP99.3
6
Annotation Error DetectionSI-Forex
AP89.7
6
Annotation Error DetectionGUM
AP98.5
6
Showing 8 of 8 rows

Other info

Code

Follow for update