Progressive Neural Networks
About
Learning to solve complex sequences of tasks--while both leveraging transfer and avoiding catastrophic forgetting--remains a key obstacle to achieving human-level intelligence. The progressive networks approach represents a step forward in this direction: they are immune to forgetting and can leverage prior knowledge via lateral connections to previously learned features. We evaluate this architecture extensively on a wide variety of reinforcement learning tasks (Atari and 3D maze games), and show that it outperforms common baselines based on pretraining and finetuning. Using a novel sensitivity measure, we demonstrate that transfer occurs at both low-level sensory and high-level control layers of the learned policy.
Andrei A. Rusu, Neil C. Rabinowitz, Guillaume Desjardins, Hubert Soyer, James Kirkpatrick, Koray Kavukcuoglu, Razvan Pascanu, Raia Hadsell• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-100 (test) | Accuracy67.2 | 3518 | |
| Image Classification | SVHN (test) | Accuracy96.8 | 362 | |
| Continual Learning | Sequential MNIST | Avg Acc99.23 | 149 | |
| Image Classification | ImageNet (val) | Accuracy76.16 | 115 | |
| Continual Learning | CIFAR100 Split | Average Per-Task Accuracy59.2 | 85 | |
| Image Classification | Stanford Cars (val) | Accuracy89.21 | 56 | |
| Continual Learning | Permuted MNIST | Mean Test Accuracy93.5 | 44 | |
| Image Classification | S-CIFAR-10 Task-IL | Accuracy95.13 | 33 | |
| Task-Incremental Learning | CIFAR100 (test) | Accuracy54.9 | 31 | |
| Image Classification | S-CIFAR-10 | Task-IL Accuracy95.13 | 27 |
Showing 10 of 36 rows