Aguvis: Unified Pure Vision Agents for Autonomous GUI Interaction
About
Automating GUI tasks remains challenging due to reliance on textual representations, platform-specific action spaces, and limited reasoning capabilities. We introduce Aguvis, a unified vision-based framework for autonomous GUI agents that directly operates on screen images, standardizes cross-platform interactions and incorporates structured reasoning via inner monologue. To enable this, we construct Aguvis Data Collection, a large-scale dataset with multimodal grounding and reasoning annotations, and develop a two-stage training pipeline that separates GUI grounding from planning and reasoning. Experiments show that Aguvis achieves state-of-the-art performance across offline and real-world online benchmarks, marking the first fully autonomous vision-based GUI agent that operates without closed-source models. We open-source all datasets, models, and training recipes at https://aguvis-project.github.io to advance future research.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| GUI Grounding | ScreenSpot v2 | Avg Accuracy86 | 203 | |
| GUI Grounding | ScreenSpot Pro | Average Score22.9 | 169 | |
| GUI Agent Task | AndroidWorld | Success Rate37.1 | 104 | |
| GUI Grounding | ScreenSpot Pro | Accuracy23.6 | 77 | |
| GUI Grounding | ScreenSpot | Avg Acc89.2 | 76 | |
| Mobile Task Automation | AndroidWorld (test) | Average Success Rate0.371 | 75 | |
| GUI Grounding | OSWorld-G | Average Score38.7 | 74 | |
| GUI Action Execution | GUI-EDA | Acoustic Score (COMSOL)53 | 60 | |
| GUI Grounding | OSWorld-G (test) | Element Accuracy41.2 | 52 | |
| Mobile GUI Automation | GUI-Odyssey | Success Rate (SR)13.5 | 50 |