KronA: Parameter Efficient Tuning with Kronecker Adapter
About
Fine-tuning a Pre-trained Language Model (PLM) on a specific downstream task has been a well-known paradigm in Natural Language Processing. However, with the ever-growing size of PLMs, training the entire model on several downstream tasks becomes very expensive and resource-hungry. Recently, different Parameter Efficient Tuning (PET) techniques are proposed to improve the efficiency of fine-tuning PLMs. One popular category of PET methods is the low-rank adaptation methods which insert learnable truncated SVD modules into the original model either sequentially or in parallel. However, low-rank decomposition suffers from limited representation power. In this work, we address this problem using the Kronecker product instead of the low-rank representation. We introduce KronA, a Kronecker product-based adapter module for efficient fine-tuning of Transformer-based PLMs. We apply the proposed methods for fine-tuning T5 on the GLUE benchmark to show that incorporating the Kronecker-based modules can outperform state-of-the-art PET methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | SQuAD 2.0 | F181.96 | 190 | |
| Summarization | Xsum | ROUGE-216.14 | 108 | |
| Question Answering | SQuAD v1.1 | F186.45 | 79 | |
| Summarization | CNN Daily Mail | ROUGE-140.83 | 67 | |
| Commonsense Reasoning | Commonsense Reasoning (BoolQ, PIQA, SIQA, HellaS., WinoG., ARC-e, ARC-c, OBQA) | BoolQ Accuracy72.9 | 61 | |
| Reading Comprehension | DROP (test) | F1 Score58.5 | 61 | |
| MRI Image Generation | ADNI (evaluation) | FID11.594 | 12 | |
| MRI Image Generation | PPMI (evaluation set) | FID15.857 | 12 | |
| MRI Image Generation | BraTS 2021 (evaluation set) | FID3.399 | 12 | |
| Concept Retrieval | ELSST concept retrieval synthetic (test) | MRR95.5 | 7 |