Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

AdaRubric: Task-Adaptive Rubrics for LLM Agent Evaluation

About

LLM-as-Judge evaluation fails agent tasks because a fixed rubric cannot capture what matters for this task: code debugging demands Correctness and Error Handling; web navigation demands Goal Alignment and Action Efficiency. We present ADARUBRIC, which closes this gap by generating task-specific evaluation rubrics on the fly from task descriptions, scoring trajectories step-by-step with confidence-weighted per-dimension feedback, and filtering preference pairs with the novel DimensionAwareFilter - a provably necessary condition for preventing high-scoring dimensions from masking dimension-level failures. On WebArena and ToolBench, ADARUBRIC achieves Pearson r=0.79 human correlation (+0.16 over the best static baseline) with deployment-grade reliability (Krippendorff's $\alpha$=0.83). DPO agents trained on ADARUBRIC preference pairs gain +6.8 to +8.5 pp task success over Prometheus across three benchmarks; gains transfer to SWE-bench code repair (+4.9 pp) and accelerate PPO convergence by +6.6 pp at 5K steps - both without any rubric engineering. Code: https://github.com/alphadl/AdaRubrics.

Liang Ding• 2026

Related benchmarks

TaskDatasetResultRank
Web Agent Task SuccessWebArena
Task Success Rate (TSR)27.8
12
Human CorrelationWebArena
Pearson Correlation Coefficient (r)0.79
8
Human CorrelationToolBench
Pearson r0.74
8
Human CorrelationAgentBench
Pearson r0.77
8
Success RateAgentBench
Success Rate34.1
8
Task Completion RateToolBench
Task Completion Rate (TCR)37.8
8
Web Agent NavigationWebArena
Success Rate27.8
8
Evaluation ReliabilityWeb Automation (WA)
Krippendorff's Alpha0.85
6
Evaluation ReliabilityToolBench (TB)
Krippendorff's Alpha0.82
6
Multimodal Agent EvaluationVisualWebArena
Pearson r0.76
6
Showing 10 of 12 rows

Other info

Follow for update