Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Gradient Projection Memory for Continual Learning

About

The ability to learn continually without forgetting the past tasks is a desired attribute for artificial learning systems. Existing approaches to enable such learning in artificial neural networks usually rely on network growth, importance based weight update or replay of old data from the memory. In contrast, we propose a novel approach where a neural network learns new tasks by taking gradient steps in the orthogonal direction to the gradient subspaces deemed important for the past tasks. We find the bases of these subspaces by analyzing network representations (activations) after learning each task with Singular Value Decomposition (SVD) in a single shot manner and store them in the memory as Gradient Projection Memory (GPM). With qualitative and quantitative analyses, we show that such orthogonal gradient descent induces minimum to no interference with the past tasks, thereby mitigates forgetting. We evaluate our algorithm on diverse image classification datasets with short and long sequences of tasks and report better or on-par performance compared to the state-of-the-art approaches.

Gobinda Saha, Isha Garg, Kaushik Roy• 2021

Related benchmarks

TaskDatasetResultRank
Language UnderstandingMMLU
Accuracy15.45
825
ReasoningBBH--
672
Physical Commonsense ReasoningPIQA
Accuracy53.48
572
Continual Image ClassificationMiniImageNet Split
Accuracy69.46
42
Continual LearningCIFAR-100 (10-split)
ACC72.48
42
Exemplar-Free Class-Incremental LearningCIFAR-100 Big start
Average Incremental Accuracy (Aavg)41.51
39
Continual Image ClassificationCIFAR100 Split
Accuracy79.58
30
Continual Image Classification5-Datasets
Accuracy (%)91.54
23
Continual LearningOL-CIFAR100 (Tasks 0-6)
Accuracy (%)71.62
23
Continual LearningMNIST permuted
AT93.91
19
Showing 10 of 24 rows

Other info

Follow for update