MiMo: Unlocking the Reasoning Potential of Language Model -- From Pretraining to Posttraining
About
We present MiMo-7B, a large language model born for reasoning tasks, with optimization across both pre-training and post-training stages. During pre-training, we enhance the data preprocessing pipeline and employ a three-stage data mixing strategy to strengthen the base model's reasoning potential. MiMo-7B-Base is pre-trained on 25 trillion tokens, with additional Multi-Token Prediction objective for enhanced performance and accelerated inference speed. During post-training, we curate a dataset of 130K verifiable mathematics and programming problems for reinforcement learning, integrating a test-difficulty-driven code-reward scheme to alleviate sparse-reward issues and employing strategic data resampling to stabilize training. Extensive evaluations show that MiMo-7B-Base possesses exceptional reasoning potential, outperforming even much larger 32B models. The final RL-tuned model, MiMo-7B-RL, achieves superior performance on mathematics, code and general reasoning tasks, surpassing the performance of OpenAI o1-mini. The model checkpoints are available at https://github.com/xiaomimimo/MiMo.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Commonsense Reasoning | HellaSwag | -- | 1460 | |
| Code Generation | HumanEval | -- | 850 | |
| Commonsense Reasoning | WinoGrande | Accuracy85.9 | 776 | |
| Mathematical Reasoning | MATH | Accuracy71 | 643 | |
| Code Generation | HumanEval+ | -- | 189 | |
| Mathematical Reasoning | GSM8K | Math Score92.3 | 171 | |
| Visual Reasoning | V*Bench | Accuracy81.7 | 58 | |
| Chinese Language Understanding | C-Eval | Accuracy91 | 47 | |
| Code Generation | MBPP+ | Score72.2 | 43 | |
| General Knowledge | MMLU-Pro | MMLU-Pro General Knowledge Score73.2 | 38 |