Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Visual-Language Prompt Tuning with Knowledge-guided Context Optimization

About

Prompt tuning is an effective way to adapt the pre-trained visual-language model (VLM) to the downstream task using task-related textual tokens. Representative CoOp-based work combines the learnable textual tokens with the class tokens to obtain specific textual knowledge. However, the specific textual knowledge is the worse generalization to the unseen classes because it forgets the essential general textual knowledge having a strong generalization ability. To tackle this issue, we introduce a novel Knowledge-guided Context Optimization (KgCoOp) to enhance the generalization ability of the learnable prompt for unseen classes. The key insight of KgCoOp is that forgetting about essential knowledge can be alleviated by reducing the discrepancy between the learnable prompt and the hand-crafted prompt. Especially, KgCoOp minimizes the discrepancy between the textual embeddings generated by learned prompts and the hand-crafted prompts. Finally, adding the KgCoOp upon the contrastive loss can make a discriminative prompt for both seen and unseen tasks. Extensive evaluation of several benchmarks demonstrates that the proposed Knowledge-guided Context Optimization is an efficient method for prompt tuning, \emph{i.e.,} achieves better performance with less training time.

Hantao Yao, Rui Zhang, Changsheng Xu• 2023

Related benchmarks

TaskDatasetResultRank
Image ClassificationEuroSAT
Accuracy46.04
497
Image ClassificationFood-101
Accuracy86.36
494
Image ClassificationDTD
Accuracy46.35
487
Image ClassificationFlowers102
Accuracy95.62
478
Image ClassificationStanford Cars
Accuracy65.41
477
Image ClassificationSUN397
Accuracy66.16
425
Image ClassificationDTD
Accuracy69.52
419
Image ClassificationUCF101
Top-1 Acc83.72
404
Image ClassificationImageNet
Top-1 Accuracy70.66
324
Image ClassificationFood101--
309
Showing 10 of 230 rows
...

Other info

Follow for update