Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Near OOD Detection for Vision-Language Prompt Learning with Contrastive Logit Score

About

Prompt learning has emerged as an efficient and effective method for fine-tuning vision-language models such as CLIP. While many studies have explored generalisation abilities of these models in few-shot classification tasks and a few studies have addressed far out-of-distribution (OOD) of the models, their potential for addressing near OOD detection remains underexplored. Existing methods either require training from scratch, need fine-tuning, or are not designed for vision-language prompt learning. To address this, we introduce the Contrastive Logit Score (CLS), a novel post-hoc, plug-and-play scoring function. CLS significantly improves near OOD detection of pre-trained vision-language prompt learning methods without modifying their model architectures or requiring retraining. Our method achieves up to an 11.67% improvement in AUROC for near OOD detection with minimal computational overhead. Extensive evaluations validate the effectiveness, efficiency, and generalisability of our approach. Our code is available at https://github.com/davidmcjung/near-OOD-prompt-learning.

Myong Chol Jung, Joanna Dipnall, Belinda Gabbe, He Zhao• 2024

Related benchmarks

TaskDatasetResultRank
Out-of-Distribution DetectionPlaces
FPR9511.05
142
OOD DetectionPlaces (OOD)
AUROC96.82
93
OOD DetectionImageNet
AUROC95.88
77
OOD DetectionCIFAR-100
AUROC86.24
66
Far Out-of-Distribution DetectioniNaturalist
FPR955.76
64
Anomaly DetectionDTD
AUROC73.59
55
OOD DetectioniNaturalist
AUROC98.68
52
Error detectionFlowers102
AuROC95.69
46
Error detectionEuroSAT
AuROC79.52
46
Far OOD detectionTexture
AUROC96.3
34
Showing 10 of 31 rows

Other info

Follow for update