PaCoRe: Learning to Scale Test-Time Compute with Parallel Coordinated Reasoning
About
We introduce Parallel Coordinated Reasoning (PaCoRe), a training-and-inference framework designed to overcome a central limitation of contemporary language models: their inability to scale test-time compute (TTC) far beyond sequential reasoning under a fixed context window. PaCoRe departs from the traditional sequential paradigm by driving TTC through massive parallel exploration coordinated via a message-passing architecture in multiple rounds. Each round launches many parallel reasoning trajectories, compacts their findings into context-bounded messages, and synthesizes these messages to guide the next round and ultimately produce the final answer. Trained end-to-end with large-scale, outcome-based reinforcement learning, the model masters the synthesis abilities required by PaCoRe and scales to multi-million-token effective TTC without exceeding context limits. The approach yields strong improvements across diverse domains, and notably pushes reasoning beyond frontier systems in mathematics: an 8B model reaches 94.5% on HMMT 2025, surpassing GPT-5's 93.2% by scaling effective TTC to roughly two million tokens. We open-source model checkpoints, training data, and the full inference pipeline to accelerate follow-up work.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mathematical Reasoning | AIME 25 | Accuracy73.3 | 201 | |
| Mathematical Reasoning | AIME 24 | AIME 24 Accuracy76.7 | 84 | |
| Mathematical Reasoning | HMMT25 | Accuracy60 | 78 | |
| Mathematical Reasoning | BRUMO25 | Accuracy80.7 | 37 | |
| Mathematical Reasoning | HMMT 24 | Pass@173.1 | 18 | |
| Mathematical Reasoning | HLE Math-text | Pass@158.1 | 12 | |
| Mathematical Reasoning | HMMT25 | Pass@169.6 | 10 | |
| Mathematical Reasoning | HMMT 24 | Iteration 0 Score26.8 | 10 | |
| Mathematical Reasoning | HMMT25 | It0 Score33.2 | 10 |