DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
About
Mathematical reasoning poses a significant challenge for language models due to its complex and structured nature. In this paper, we introduce DeepSeekMath 7B, which continues pre-training DeepSeek-Coder-Base-v1.5 7B with 120B math-related tokens sourced from Common Crawl, together with natural language and code data. DeepSeekMath 7B has achieved an impressive score of 51.7% on the competition-level MATH benchmark without relying on external toolkits and voting techniques, approaching the performance level of Gemini-Ultra and GPT-4. Self-consistency over 64 samples from DeepSeekMath 7B achieves 60.9% on MATH. The mathematical reasoning capability of DeepSeekMath is attributed to two key factors: First, we harness the significant potential of publicly available web data through a meticulously engineered data selection pipeline. Second, we introduce Group Relative Policy Optimization (GRPO), a variant of Proximal Policy Optimization (PPO), that enhances mathematical reasoning abilities while concurrently optimizing the memory usage of PPO.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mathematical Reasoning | GSM8K | Accuracy92.8 | 983 | |
| Code Generation | HumanEval | Pass@194.71 | 850 | |
| Mathematical Reasoning | GSM8K (test) | Accuracy86.7 | 797 | |
| Mathematical Reasoning | GSM8K (test) | Accuracy88.2 | 751 | |
| Question Answering | ARC Challenge | Accuracy52 | 749 | |
| Mathematical Reasoning | MATH | Accuracy87.8 | 643 | |
| Mathematical Reasoning | MATH | Accuracy87.8 | 535 | |
| Reasoning | BBH | Accuracy86.1 | 507 | |
| Code Generation | HumanEval (test) | Pass@175.61 | 444 | |
| Mathematical Reasoning | MATH (test) | Overall Accuracy58.8 | 433 |