REVEALER: Reinforcement-Guided Visual Reasoning for Element-Level Text-Image Alignment Evaluation
About
Evaluating the alignment between textual prompts and generated images is critical for ensuring the reliability and usability of text-to-image (T2I) models. However, most existing evaluation methods rely on coarse-grained metrics or static QA pipelines, which lack fine-grained interpretability and struggle to reflect human preferences. To address this, we propose REVEALER, a unified framework for element-level alignment evaluation based on reinforcement-guided visual reasoning. Adopting a structured "grounding-reasoning-conclusion" paradigm, our method enables Multimodal Large Language Models (MLLMs) to explicitly localize semantic elements and derive interpretable alignment judgments. We optimize the model via Group Relative Policy Optimization(GRPO) using a composite reward function that incorporates structural format, grounding accuracy, and alignment fidelity. Extensive experiments across four benchmarks-EvalMuse-40K, RichHF, MHaluBench, and GenAI-Bench-demonstrate that REVEALER achieves state-of-the-art performance. Our approach consistently outperforms both strong proprietary models and supervised baselines while demonstrating superior inference efficiency compared to existing iterative visual reasoning methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| element-level text-to-image alignment evaluation | EvalMuse-40K | SRCC72.3 | 17 | |
| element-level text-to-image alignment evaluation | RichHF | SRCC73.3 | 17 | |
| element-level text-to-image alignment evaluation | MHaluBench | SRCC72.7 | 17 | |
| element-level text-to-image alignment evaluation | GenAI-Bench | SRCC0.749 | 17 | |
| Visual Reasoning | RichHF | SRCC70.8 | 5 | |
| Visual Reasoning | MHaluBench | SRCC70.6 | 5 | |
| Visual Reasoning | GenAI-Bench | SRCC74.4 | 5 |