Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Dark Experience for General Continual Learning: a Strong, Simple Baseline

About

Continual Learning has inspired a plethora of approaches and evaluation settings; however, the majority of them overlooks the properties of a practical scenario, where the data stream cannot be shaped as a sequence of tasks and offline training is not viable. We work towards General Continual Learning (GCL), where task boundaries blur and the domain and class distributions shift either gradually or suddenly. We address it through mixing rehearsal with knowledge distillation and regularization; our simple baseline, Dark Experience Replay, matches the network's logits sampled throughout the optimization trajectory, thus promoting consistency with its past. By conducting an extensive analysis on both standard benchmarks and a novel GCL evaluation setting (MNIST-360), we show that such a seemingly simple baseline outperforms consolidated approaches and leverages limited resources. We further explore the generalization capabilities of our objective, showing its regularization being beneficial beyond mere performance.

Pietro Buzzega, Matteo Boschini, Angelo Porrello, Davide Abati, Simone Calderara• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)--
3518
Image ClassificationCIFAR-10 (test)--
3381
Image ClassificationCIFAR-10--
507
Depth EstimationNYU v2 (test)--
432
Time Series ForecastingETTm2--
382
Image ClassificationTiny ImageNet (test)
Accuracy70.28
362
Semantic segmentationNYU v2 (test)
mIoU27.12
282
Class-incremental learningCIFAR-100--
248
Surface Normal EstimationNYU v2 (test)
Mean Angle Distance (MAD)31.25
224
Continual LearningSequential MNIST
Avg Acc99.93
149
Showing 10 of 337 rows
...

Other info

Code

Follow for update