Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Episodic Memory in Lifelong Language Learning

About

We introduce a lifelong language learning setup where a model needs to learn from a stream of text examples without any dataset identifier. We propose an episodic memory model that performs sparse experience replay and local adaptation to mitigate catastrophic forgetting in this setup. Experiments on text classification and question answering demonstrate the complementary benefits of sparse experience replay and local adaptation to allow the model to continuously learn from new datasets. We also show that the space complexity of the episodic memory module can be reduced significantly (~50-90%) by randomly choosing which examples to store in memory with a minimal decrease in performance. We consider an episodic memory component as a crucial building block of general linguistic intelligence and see our model as a first step in that direction.

Cyprien de Masson d'Autume, Sebastian Ruder, Lingpeng Kong, Dani Yogatama• 2019

Related benchmarks

TaskDatasetResultRank
Text ClassificationYahoo! Answers (test)--
133
Incremental LearningTinyImageNet
Avg Incremental Accuracy8.49
83
Text ClassificationAGNews, Amazon, DBPedia, Yahoo, and Yelp (test)
Exact Match (EM)76.7
55
Text ClassificationYelp (test)--
55
Continual LearningLarge Number of Tasks
Average Performance7.4
50
Continual LearningStandard CL Benchmark
BWT (Avg Order 1-3)57.8
38
Image ClassificationS-MNIST (test)
Average Accuracy99.2
18
Text ClassificationAGNews, Yelp, Amazon, DBPedia, Yahoo (last epoch of last task)
EM Score74.9
15
Image ClassificationS-CIFAR100 (test)
Average Accuracy60.1
14
Image ClassificationS-TinyImageNet (test)
Average Accuracy35.6
14
Showing 10 of 12 rows

Other info

Follow for update