T-pro 2.0: An Efficient Russian Hybrid-Reasoning Model and Playground
About
We introduce T-pro 2.0, an open-weight Russian LLM for hybrid reasoning and efficient inference. The model supports direct answering and reasoning-trace generation, using a Cyrillic-dense tokenizer and an adapted EAGLE speculative-decoding pipeline to reduce latency. To enable reproducible and extensible research, we release the model weights, the T-Wix 500k instruction corpus, the T-Math reasoning benchmark, and the EAGLE weights on Hugging Face. These resources allow users to study Russian-language reasoning and to extend or adapt both the model and the inference pipeline. A public web demo exposes reasoning and non-reasoning modes and illustrates the speedups achieved by our inference stack across domains. T-pro 2.0 thus serves as an accessible open system for building and evaluating efficient, practical Russian LLM applications.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Advanced Reasoning | Vikhr Physics | Accuracy51 | 11 | |
| Advanced Reasoning | T-Math | Accuracy54.1 | 11 | |
| Advanced Reasoning | ruAIME 2025 | Accuracy64.6 | 11 | |
| Advanced Reasoning | ruMATH-500 | Accuracy94 | 11 | |
| Coding Reasoning | ruLCB | Accuracy0.563 | 11 | |
| Advanced Reasoning | ruAIME 2024 | Accuracy70.4 | 11 | |
| Advanced Reasoning | Vikhr Math | Accuracy79.9 | 11 | |
| Advanced Reasoning | ruGPQA Diamond | Accuracy0.591 | 11 | |
| Mathematical Reasoning | T-Math (full) | Pass@154 | 6 | |
| Reward Model Evaluation | Arena-Hard RU | Best@8 Score92.69 | 5 |