Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference

About

Lack of performance when it comes to continual learning over non-stationary distributions of data remains a major challenge in scaling neural network learning to more human realistic settings. In this work we propose a new conceptualization of the continual learning problem in terms of a temporally symmetric trade-off between transfer and interference that can be optimized by enforcing gradient alignment across examples. We then propose a new algorithm, Meta-Experience Replay (MER), that directly exploits this view by combining experience replay with optimization based meta-learning. This method learns parameters that make interference based on future gradients less likely and transfer based on future gradients more likely. We conduct experiments across continual lifelong supervised learning benchmarks and non-stationary reinforcement learning environments demonstrating that our approach consistently outperforms recently proposed baselines for continual learning. Our experiments show that the gap between the performance of MER and baseline algorithms grows both as the environment gets more non-stationary and as the fraction of the total experiences stored gets smaller.

Matthew Riemer, Ignacio Cases, Robert Ajemian, Miao Liu, Irina Rish, Yuhai Tu, Gerald Tesauro• 2018

Related benchmarks

TaskDatasetResultRank
Continual LearningSequential MNIST
Avg Acc99.93
149
Medical Image SegmentationLA
Dice87.37
97
Continual LearningCamelyon-TCGA
AACC49.4
64
Class-incremental learningSplit CIFAR-100 (10-task)
CAA5.4
41
Class-incremental learningCIFAR-10 Sequential
FAA85.91
39
Class-Incremental Continual LearningCIFAR-10 Sequential
Forgetting17.15
39
Task-Incremental LearningCIFAR10 Sequential
Final Average Accuracy93.62
39
Class-incremental learningSequential MNIST
Forgetting1.46
33
Incremental Task Learning (ITL)Permuted MNIST (test)
Retained Accuracy97.15
32
Incremental Task Learning (ITL)split-MNIST (test)
Retained Accuracy97.12
32
Showing 10 of 91 rows
...

Other info

Follow for update