Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Label Propagation for Zero-shot Classification with Vision-Language Models

About

Vision-Language Models (VLMs) have demonstrated impressive performance on zero-shot classification, i.e. classification when provided merely with a list of class names. In this paper, we tackle the case of zero-shot classification in the presence of unlabeled data. We leverage the graph structure of the unlabeled data and introduce ZLaP, a method based on label propagation (LP) that utilizes geodesic distances for classification. We tailor LP to graphs containing both text and image features and further propose an efficient method for performing inductive inference based on a dual solution and a sparsification step. We perform extensive experiments to evaluate the effectiveness of our method on 14 common datasets and show that ZLaP outperforms the latest related works. Code: https://github.com/vladan-stojnic/ZLaP

Vladan Stojni\'c, Yannis Kalantidis, Giorgos Tolias• 2024

Related benchmarks

TaskDatasetResultRank
Image ClassificationEuroSAT
Accuracy60.9
497
Image ClassificationDTD
Accuracy51.8
487
Image ClassificationUCF101
Top-1 Acc77.7
404
ClassificationCars
Accuracy72.1
314
Image ClassificationCUB
Accuracy64.1
249
Image ClassificationFGVCAircraft
Accuracy28.4
225
Image ClassificationPets
Accuracy92.8
204
Image ClassificationFlowers
Accuracy73.4
127
Image ClassificationCaltech
Accuracy91.8
98
Image ClassificationFood
Accuracy87.9
92
Showing 10 of 14 rows

Other info

Code

Follow for update