SuS-X: Training-Free Name-Only Transfer of Vision-Language Models
About
Contrastive Language-Image Pre-training (CLIP) has emerged as a simple yet effective way to train large-scale vision-language models. CLIP demonstrates impressive zero-shot classification and retrieval on diverse downstream tasks. However, to leverage its full potential, fine-tuning still appears to be necessary. Fine-tuning the entire CLIP model can be resource-intensive and unstable. Moreover, recent methods that aim to circumvent this need for fine-tuning still require access to images from the target distribution. In this paper, we pursue a different approach and explore the regime of training-free "name-only transfer" in which the only knowledge we possess about the downstream task comprises the names of downstream target categories. We propose a novel method, SuS-X, consisting of two key building blocks -- SuS and TIP-X, that requires neither intensive fine-tuning nor costly labelled data. SuS-X achieves state-of-the-art zero-shot classification results on 19 benchmark datasets. We further show the utility of TIP-X in the training-free few-shot setting, where we again achieve state-of-the-art results over strong training-free baselines. Code is available at https://github.com/vishaal27/SuS-X.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | Stanford Cars | Accuracy67.3 | 635 | |
| Image Classification | ImageNet-1K | Top-1 Acc69.88 | 600 | |
| Image Classification | EuroSAT | Accuracy73.1 | 569 | |
| Image Classification | Flowers102 | Accuracy90.3 | 558 | |
| Image Classification | Food-101 | -- | 542 | |
| Image Classification | DTD | -- | 542 | |
| Image Classification | Food101 | Accuracy77.9 | 457 | |
| Image Classification | UCF101 | Top-1 Acc66.59 | 455 | |
| Image Classification | SUN397 | Accuracy68 | 441 | |
| Image Classification | SUN397 | -- | 425 |