Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Progressive Prompts: Continual Learning for Language Models

About

We introduce Progressive Prompts - a simple and efficient approach for continual learning in language models. Our method allows forward transfer and resists catastrophic forgetting, without relying on data replay or a large number of task-specific parameters. Progressive Prompts learns a new soft prompt for each task and sequentially concatenates it with the previously learned prompts, while keeping the base model frozen. Experiments on standard continual learning benchmarks show that our approach outperforms state-of-the-art methods, with an improvement >20% in average test accuracy over the previous best-preforming method on T5 model. We also explore a more challenging continual learning setup with longer sequences of tasks and show that Progressive Prompts significantly outperforms prior methods.

Anastasia Razdaibiedina, Yuning Mao, Rui Hou, Madian Khabsa, Mike Lewis, Amjad Almahairi• 2023

Related benchmarks

TaskDatasetResultRank
Language UnderstandingMMLU
Accuracy30.58
756
ReasoningBBH--
507
Physical Commonsense ReasoningPIQA
Accuracy53.05
329
Text ClassificationAGNews, Amazon, DBPedia, Yahoo, and Yelp (test)
Exact Match (EM)79.9
55
Continual LearningLarge Number of Tasks
Average Performance78
50
Continual LearningStandard CL Benchmark
BWT (Avg Order 1-3)76.31
38
Continual LearningLong Sequence (test)
AP7.98
15
Continual LearningSuperNI (test)
AP3.34
13
Language ModelingGeneral Knowledge Base Model Knowledge
GEN Loss-20.9
10
Continual LearningNumGLUE-cm
Average Accuracy25.3
9
Showing 10 of 23 rows

Other info

Follow for update