Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

RADIOv2.5: Improved Baselines for Agglomerative Vision Foundation Models

About

Agglomerative models have recently emerged as a powerful approach to training vision foundation models, leveraging multi-teacher distillation from existing models such as CLIP, DINO, and SAM. This strategy enables the efficient creation of robust models, combining the strengths of individual teachers while significantly reducing computational and resource demands. In this paper, we thoroughly analyze state-of-the-art agglomerative models, identifying critical challenges including resolution mode shifts, teacher imbalance, idiosyncratic teacher artifacts, and an excessive number of output tokens. To address these issues, we propose several novel solutions: multi-resolution training, mosaic augmentation, and improved balancing of teacher loss functions. Specifically, in the context of Vision Language Models, we introduce a token compression technique to maintain high-resolution information within a fixed token count. We release our top-performing variants at multiple scales (-B, -L, -H, and -g), along with inference code and pretrained weights

Greg Heinrich, Mike Ranzinger, Hongxu (Danny) Yin, Yao Lu, Jan Kautz, Andrew Tao, Bryan Catanzaro, Pavlo Molchanov• 2024

Related benchmarks

TaskDatasetResultRank
Semantic segmentationADE20K (val)
mIoU53.91
2888
Object Hallucination EvaluationPOPE--
1455
Visual Question AnsweringGQA
Accuracy65.8
1249
Image ClassificationImageNet-1K
Top-1 Acc85.81
1239
Semantic segmentationADE20K
mIoU54.56
1024
Semantic segmentationCityscapes
mIoU64.11
658
Multimodal EvaluationMME--
658
Image ClassificationStanford Cars
Accuracy82
635
Image ClassificationImageNet-1K--
600
Image ClassificationFood-101
Accuracy89.4
542
Showing 10 of 79 rows
...

Other info

Follow for update