AnomalyVFM -- Transforming Vision Foundation Models into Zero-Shot Anomaly Detectors
About
Zero-shot anomaly detection aims to detect and localise abnormal regions in the image without access to any in-domain training images. While recent approaches leverage vision-language models (VLMs), such as CLIP, to transfer high-level concept knowledge, methods based on purely vision foundation models (VFMs), like DINOv2, have lagged behind in performance. We argue that this gap stems from two practical issues: (i) limited diversity in existing auxiliary anomaly detection datasets and (ii) overly shallow VFM adaptation strategies. To address both challenges, we propose AnomalyVFM, a general and effective framework that turns any pretrained VFM into a strong zero-shot anomaly detector. Our approach combines a robust three-stage synthetic dataset generation scheme with a parameter-efficient adaptation mechanism, utilising low-rank feature adapters and a confidence-weighted pixel loss. Together, these components enable modern VFMs to substantially outperform current state-of-the-art methods. More specifically, with RADIO as a backbone, AnomalyVFM achieves an average image-level AUROC of 94.1% across 9 diverse datasets, surpassing previous methods by significant 3.3 percentage points. Project Page: https://maticfuc.github.io/anomaly_vfm/
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Anomaly Localization | MVTec AD | Pixel AUROC92.7 | 369 | |
| Anomaly Detection | MVTec-AD (test) | I-AUROC98.2 | 226 | |
| Anomaly Detection | VisA | AUROC93.6 | 199 | |
| Anomaly Detection | VisA (test) | I-AUROC94.5 | 91 | |
| Anomaly Detection | KSDD | AUROC0.925 | 40 | |
| Anomaly Detection | DTD | AUROC99.4 | 28 | |
| Image-level Anomaly Detection | HeadCT | AUROC94.8 | 24 | |
| Anomaly Detection | BTAD | AUROC96 | 20 | |
| Anomaly Localization | DTD | AUROC99.4 | 19 | |
| Anomaly Localization | KSDD | AUROC99 | 19 |