Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Continual Learning with Deep Generative Replay

About

Attempts to train a comprehensive artificial intelligence capable of solving multiple tasks have been impeded by a chronic problem called catastrophic forgetting. Although simply replaying all previous data alleviates the problem, it requires large memory and even worse, often infeasible in real world applications where the access to past data is limited. Inspired by the generative nature of hippocampus as a short-term memory system in primate brain, we propose the Deep Generative Replay, a novel framework with a cooperative dual model architecture consisting of a deep generative model ("generator") and a task solving model ("solver"). With only these two models, training data for previous tasks can easily be sampled and interleaved with those for a new task. We test our methods in several sequential learning settings involving image classification tasks.

Hanul Shin, Jung Kwon Lee, Jaehong Kim, Jiwon Kim• 2017

Related benchmarks

TaskDatasetResultRank
Class-incremental learningCIFAR-100 10 (test)
Average Top-1 Accuracy8.1
75
Image ClassificationSplit MNIST
Average Accuracy91.79
49
Incremental Task Learning (ITL)split-MNIST (test)
Retained Accuracy99.47
32
Incremental Task Learning (ITL)Permuted MNIST (test)
Retained Accuracy92.52
32
Class-incremental learningCIFAR-100 5 tasks (test)
AN14.4
20
Class-incremental learningCIFAR-100 20 tasks (test)
Average Novelty4.1
20
Continual LearningPMNIST (test)
Accuracy94.9
17
Incremental Domain Learning (IDL)split-MNIST (test)
Retained Accuracy95.74
16
Incremental Domain Learning (IDL)Permuted MNIST (test)
Retained Accuracy95.09
16
Incremental LearningMNIST
Average Accuracy0.9124
10
Showing 10 of 10 rows

Other info

Follow for update