Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

VLMInferSlow: Evaluating the Efficiency Robustness of Large Vision-Language Models as a Service

About

Vision-Language Models (VLMs) have demonstrated great potential in real-world applications. While existing research primarily focuses on improving their accuracy, the efficiency remains underexplored. Given the real-time demands of many applications and the high inference overhead of VLMs, efficiency robustness is a critical issue. However, previous studies evaluate efficiency robustness under unrealistic assumptions, requiring access to the model architecture and parameters -- an impractical scenario in ML-as-a-service settings, where VLMs are deployed via inference APIs. To address this gap, we propose VLMInferSlow, a novel approach for evaluating VLM efficiency robustness in a realistic black-box setting. VLMInferSlow incorporates fine-grained efficiency modeling tailored to VLM inference and leverages zero-order optimization to search for adversarial examples. Experimental results show that VLMInferSlow generates adversarial images with imperceptible perturbations, increasing the computational cost by up to 128.47%. We hope this research raises the community's awareness about the efficiency robustness of VLMs.

Xiasi Wang, Tianliang Yao, Simin Chen, Runqi Wang, Lei YE, Kuofeng Gao, Yi Huang, Yuan Yao• 2025

Related benchmarks

TaskDatasetResultRank
Inference EfficiencyMS-COCO
Sequence Length Delta51.96
20
Inference EfficiencyImageNet-1K
Inference Length47.39
20
Showing 2 of 2 rows

Other info

Code

Follow for update