Large-Scale Universal Defect Generation: Foundation Models and Datasets
About
Existing defect/anomaly generation methods often rely on few-shot learning, which overfits to specific defect categories due to the lack of large-scale paired defect editing data. This issue is aggravated by substantial variations in defect scale and morphology, resulting in limited generalization, degraded realism, and category consistency. We address these challenges by introducing UDG, a large-scale dataset of 300K normal-abnormal-mask-caption quadruplets spanning diverse domains, and by presenting UniDG, a universal defect generation foundation model that supports both reference-based defect generation and text instruction-based defect editing without per-category fine-tuning. UniDG performs Defect-Context Editing via adaptive defect cropping and structured diptych input format, and fuses reference and target conditions through MM-DiT multimodal attention. A two-stage training strategy, Diversity-SFT followed by Consistency-RFT, further improves diversity while enhancing realism and reference consistency. Extensive experiments on MVTec-AD and VisA show that UniDG outperforms prior few-shot anomaly generation and image insertion/editing baselines in synthesis quality and downstream single- and multi-class anomaly detection/localization. Code will be available at https://github.com/RetoFan233/UniDG.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Anomaly Localization | MVTec AD | Pixel AUROC98.8 | 513 | |
| Anomaly Detection | VisA | -- | 52 | |
| Anomaly Detection | MVTec AD | Image AUROC96.49 | 29 | |
| Anomaly Detection | MVTec AD | Img AUROC98.4 | 16 | |
| Anomaly Classification | MVTec AD | -- | 10 | |
| Semantic segmentation | MVTec AD | mIoU44.21 | 8 | |
| Anomaly Synthesis Quality | MVTec AD | IL0.29 | 8 | |
| Semantic segmentation | VisA | mIoU36.23 | 3 |