Progressive Prompts: Continual Learning for Language Models
About
We introduce Progressive Prompts - a simple and efficient approach for continual learning in language models. Our method allows forward transfer and resists catastrophic forgetting, without relying on data replay or a large number of task-specific parameters. Progressive Prompts learns a new soft prompt for each task and sequentially concatenates it with the previously learned prompts, while keeping the base model frozen. Experiments on standard continual learning benchmarks show that our approach outperforms state-of-the-art methods, with an improvement >20% in average test accuracy over the previous best-preforming method on T5 model. We also explore a more challenging continual learning setup with longer sequences of tasks and show that Progressive Prompts significantly outperforms prior methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Language Understanding | MMLU | Accuracy30.58 | 756 | |
| Reasoning | BBH | -- | 507 | |
| Physical Commonsense Reasoning | PIQA | Accuracy53.05 | 329 | |
| Text Classification | AGNews, Amazon, DBPedia, Yahoo, and Yelp (test) | Exact Match (EM)79.9 | 55 | |
| Continual Learning | Large Number of Tasks | Average Performance78 | 50 | |
| Continual Learning | Standard CL Benchmark | BWT (Avg Order 1-3)76.31 | 38 | |
| Continual Learning | Long Sequence (test) | AP7.98 | 15 | |
| Continual Learning | SuperNI (test) | AP3.34 | 13 | |
| Language Modeling | General Knowledge Base Model Knowledge | GEN Loss-20.9 | 10 | |
| Continual Learning | NumGLUE-cm | Average Accuracy25.3 | 9 |