Task Matrices: Linear Maps for Cross-Model Finetuning Transfer
About
Results in interpretability suggest that large vision and language models learn implicit linear encodings when models are biased by in-context prompting. However, the existence of similar linear representations in more general adaptation regimes has not yet been demonstrated. In this work, we develop the concept of a task matrix, a linear transformation from a base to finetuned embedding state. We demonstrate that for vision and text models and ten different datasets, a base model augmented with a task matrix achieves results surpassing linear probes, sometimes approaching finetuned levels. Our results validate the existence of cross-layer linear encodings between pretrained and finetuned architectures. Moreover, we show that a data-based approximation for such encodings is both efficient and generalizable to multiple domains. We make our implementation publicly available.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | EuroSAT | Accuracy97.1 | 497 | |
| Image Classification | Stanford Cars | Accuracy93.7 | 477 | |
| Image Classification | DTD | Accuracy82.9 | 419 | |
| Image Classification | GTSRB | Accuracy87.2 | 291 | |
| Image Classification | SUN397 | Accuracy76.7 | 246 | |
| Natural Language Inference | SNLI | Accuracy76.3 | 174 | |
| Image Classification | CIFAR100 | Average Accuracy86.7 | 121 | |
| Image Classification | MNIST | Accuracy99.03 | 48 | |
| Image Classification | SVHN | Accuracy66.7 | 30 | |
| Intent Detection | ATIS | ID Accuracy95.5 | 27 |