Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning from Noisy Labels for Entity-Centric Information Extraction

About

Recent information extraction approaches have relied on training deep neural models. However, such models can easily overfit noisy labels and suffer from performance degradation. While it is very costly to filter noisy labels in large learning resources, recent studies show that such labels take more training steps to be memorized and are more frequently forgotten than clean labels, therefore are identifiable in training. Motivated by such properties, we propose a simple co-regularization framework for entity-centric information extraction, which consists of several neural models with identical structures but different parameter initialization. These models are jointly optimized with the task-specific losses and are regularized to generate similar predictions based on an agreement loss, which prevents overfitting on noisy labels. Extensive experiments on two widely used but noisy benchmarks for information extraction, TACRED and CoNLL03, demonstrate the effectiveness of our framework. We release our code to the community for future research.

Wenxuan Zhou, Muhao Chen• 2021

Related benchmarks

TaskDatasetResultRank
Named Entity RecognitionCoNLL 2003 (test)
F1 Score94.22
539
Relation ExtractionTACRED (test)
F1 Score73
194
Named Entity RecognitionCoNLL 2003 (dev)
F1 Score97.21
40
News topic classification20 Newsgroups 20% Symmetric Noise
Accuracy83.09
24
News topic classification20 Newsgroups 20% Asymmetric Noise
Accuracy83.13
24
News topic classification20 Newsgroups 40% Symmetric Noise
Accuracy77.96
24
News topic classification20 Newsgroups 20% Instance-Dependent Noise
Accuracy83.47
24
News topic classification20 Newsgroups 40% Instance-Dependent Noise
Accuracy80.47
24
News topic classification20 Newsgroups 40% Asymmetric Noise
Accuracy73.5
24
Text ClassificationAGNews 4 classes symmetric noise e=0.4 (test)
Accuracy83.21
24
Showing 10 of 20 rows

Other info

Code

Follow for update