Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Sentence Embedding Alignment for Lifelong Relation Extraction

About

Conventional approaches to relation extraction usually require a fixed set of pre-defined relations. Such requirement is hard to meet in many real applications, especially when new data and relations are emerging incessantly and it is computationally expensive to store all data and re-train the whole model every time new data and relations come in. We formulate such a challenging problem as lifelong relation extraction and investigate memory-efficient incremental learning methods without catastrophically forgetting knowledge learned from previous tasks. We first investigate a modified version of the stochastic gradient methods with a replay memory, which surprisingly outperforms recent state-of-the-art lifelong learning methods. We further propose to improve this approach to alleviate the forgetting problem by anchoring the sentence embedding space. Specifically, we utilize an explicit alignment model to mitigate the sentence embedding distortion of the learned model when training on new data and new relations. Experiment results on multiple benchmarks show that our proposed method significantly outperforms the state-of-the-art lifelong learning approaches.

Hong Wang, Wenhan Xiong, Mo Yu, Xiaoxiao Guo, Shiyu Chang, William Yang Wang• 2019

Related benchmarks

TaskDatasetResultRank
Event ClassificationMAVEN (test dev)
Micro-F1 (Session 1)88.9
11
Event DetectionACE05 (test dev)
Micro-F1 (Session 1)59.6
8
Event DetectionMAVEN (test dev)
Micro-F1 (Session 1)62
8
Continual Relation ExtractionFewRel (incremental)
T189
8
Continual Relation ExtractionTACRED (incremental)
Performance T10.475
8
Continual Relation LearningFewRel 10-way 5-shot (test)
Accuracy (Task 1)0.9278
6
Relation ClassificationFewRel 10-way 5-shot
Metric 1 Accuracy92.78
6
Continual Relation ClassificationFewRel 10-way 5-shot (test)
Task 1 Accuracy96.35
6
Showing 8 of 8 rows

Other info

Follow for update