Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Dynamic Adversarial Reinforcement Learning for Robust Multimodal Large Language Models

About

Despite their impressive capabilities, Multimodal Large Language Models (MLLMs) exhibit perceptual fragility when confronted with visually complex scenes. This weakness stems from a reliance on finite training datasets, which are prohibitively expensive to scale and impose a ceiling on model robustness. We introduce \textbf{AOT-SFT}, a large-scale adversarial dataset for bootstrapping MLLM robustness. Building on this, we propose \textbf{AOT (Adversarial Opponent Training)}, a self-play framework that forges MLLM robustness by creating its own training data. Our method orchestrates a co-evolution between an image-editing Attacker and a Defender MLLM, where the Attacker generates a diverse and dynamic curriculum of image manipulations, forcing the Defender to adapt and improve. Extensive experiments demonstrate that AOT enhances the Defender's perceptual robustness and reduces hallucinations, establishing a scalable paradigm for training more reliable MLLMs.

Yicheng Bao, Xuhong Wang, Qiaosheng Zhang, Chaochao Lu, Xia Hu, Xin Tan• 2026

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringRealworldQA
Accuracy70.07
179
Hallucination EvaluationPOPE--
153
Multimodal UnderstandingMMMU (val)--
152
Multimodal UnderstandingSEEDBench2 Plus
Accuracy70.05
74
Hallucination assessmentHallusionBench
Answer Accuracy (aAcc)69.19
39
Multi-modal Visual CapabilityMMStar
Score61.53
29
Multi-image visual perceptionBLINK
Accuracy55.92
26
High-Resolution Multimodal UnderstandingHRBench-8K
Accuracy71.5
13
Multidisciplinary knowledge and reasoningMMMU (dev)
Score25.33
9
Perceptual RobustnessVSTAR
Overall Accuracy80.25
9
Showing 10 of 15 rows

Other info

Follow for update