Let's Roll a BiFTA: Bi-refinement for Fine-grained Text-visual Alignment in Vision-Language Models
About
Recent research has shown that aligning fine-grained text descriptions with localized image patches can significantly improve the zero-shot performance of pre-trained vision-language models (e.g., CLIP). However, we find that both fine-grained text descriptions and localized image patches often contain redundant information, making text-visual alignment less effective. In this paper, we tackle this issue from two perspectives: \emph{View Refinement} and \emph{Description refinement}, termed as \textit{\textbf{Bi}-refinement for \textbf{F}ine-grained \textbf{T}ext-visual \textbf{A}lignment} (BiFTA). \emph{View refinement} removes redundant image patches with high \emph{Intersection over Union} (IoU) ratios, resulting in more distinctive visual samples. \emph{Description refinement} removes redundant text descriptions with high pairwise cosine similarity, ensuring greater diversity in the remaining descriptions. BiFTA achieves superior zero-shot performance on 6 benchmark datasets for both ViT-based and ResNet-based CLIP, justifying the necessity to remove redundant information in visual-text alignment.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | DTD | Accuracy50.87 | 487 | |
| Image Classification | ImageNet | -- | 429 | |
| Image Classification | ImageNet 1k (test) | Top-1 Accuracy77.74 | 359 | |
| Image Classification | ImageNet (test) | Top-1 Accuracy77.82 | 291 | |
| Image Classification | CUB-200 2011 | Accuracy45.86 | 257 | |
| Image Classification | Food101 (test) | Accuracy93.97 | 87 | |
| Classification | CUB | Accuracy58.24 | 85 | |
| Classification | CUB (test) | Accuracy65.67 | 79 | |
| Image Classification | Downstream Datasets Average | Average Accuracy72.98 | 57 | |
| Classification | Food101 | -- | 51 |