GRPO-LEAD: A Difficulty-Aware Reinforcement Learning Approach for Concise Mathematical Reasoning in Language Models
About
Group Relative Policy Optimization (GRPO), which is widely adopted by R1-like reasoning models, has advanced mathematical reasoning. Nevertheless, GRPO faces challenges in reward sparsity, verbosity, and inadequate focus on problem difficulty. We propose GRPO-LEAD, enhancing GRPO with: (1) length-regularized rewards to encourage conciseness while maintaining accuracy; (2) explicit penalties for incorrect solutions to improve model precision; and (3) difficulty-aware advantage reweighting for robust generalization on challenging problems. Comprehensive evaluations demonstrate that GRPO-LEAD significantly improves reasoning accuracy, conciseness, and efficiency. Our approach achieves state-of-the-art performance for 14B-scale models, underscoring the synergy of our methods with appropriate model scale and high-quality data. Our source code, generated dataset, and models are available at https://github.com/aeroplanepaper/GRPO-LEAD.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mathematical Reasoning | CollegeMATH | -- | 161 | |
| Mathematical Reasoning | Minerva | Pass@129.14 | 138 | |
| Visual Mathematical Reasoning | MathVerse | -- | 73 | |
| Mathematical Reasoning | AIME 2024 | Pass@153.1 | 54 | |
| Mathematical Reasoning | OlympiadBench | Pass@140.6 | 39 | |
| Mathematical Reasoning | MATH 500 | Pass@189.5 | 33 | |
| Mathematical Reasoning | AIME 2025 | Pass@136.1 | 33 | |
| Mathematical Reasoning | Mathematical Reasoning Aggregate | Average Score38.26 | 18 | |
| Mathematical Reasoning | Olympiad | Score37.07 | 17 | |
| Visual Mathematical Reasoning | MathVista | Avg@30.702 | 14 |