Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Multitask Prompt Tuning Enables Parameter-Efficient Transfer Learning

About

Prompt tuning, in which a base pretrained model is adapted to each task via conditioning on learned prompt vectors, has emerged as a promising approach for efficiently adapting large language models to multiple downstream tasks. However, existing methods typically learn soft prompt vectors from scratch, and it has not been clear how to exploit the rich cross-task knowledge with prompt vectors in a multitask learning setting. We propose multitask prompt tuning (MPT), which first learns a single transferable prompt by distilling knowledge from multiple task-specific source prompts. We then learn multiplicative low rank updates to this shared prompt to efficiently adapt it to each downstream target task. Extensive experiments on 23 NLP datasets demonstrate that our proposed approach outperforms the state-of-the-art methods, including the full finetuning baseline in some cases, despite only tuning 0.035% as many task-specific parameters.

Zhen Wang, Rameswar Panda, Leonid Karlinsky, Rogerio Feris, Huan Sun, Yoon Kim• 2023

Related benchmarks

TaskDatasetResultRank
Reading ComprehensionC3
Accuracy50.64
73
Aspect-level Sentiment AnalysisCOTE BD
F1 Score85.32
34
Natural Language UnderstandingSuperGLUE
MultiRC Score74.8
22
Extractive Question AnsweringMRQA
NewsQA Score63.7
19
Natural Language UnderstandingGLUE & SuperGLUE (test)
RTE Accuracy79.4
17
Relation ExtractionFinRE
F1 Score75.96
17
Reading ComprehensionDRCD
F1 Score85.12
17
Reading ComprehensionSanWen syntactically perturbed
F1 Score90.48
17
Semantic SimilarityLCQMC
Accuracy75.26
17
Sentiment AnalysisAmazon syntactically perturbed
Accuracy62.86
17
Showing 10 of 22 rows

Other info

Follow for update