JoyAI-LLM Flash: Advancing Mid-Scale LLMs with Token Efficiency
About
We introduce JoyAI-LLM Flash, an efficient Mixture-of-Experts (MoE) language model designed to redefine the trade-off between strong performance and token efficiency in the sub-50B parameter regime. JoyAI-LLM Flash is pretrained on a massive corpus of 20 trillion tokens and further optimized through a rigorous post-training pipeline, including supervised fine-tuning (SFT), Direct Preference Optimization (DPO), and large-scale reinforcement learning (RL) across diverse environments. To improve token efficiency, JoyAI-LLM Flash strategically balances \emph{thinking} and \emph{non-thinking} cognitive modes and introduces FiberPO, a novel RL algorithm inspired by fibration theory that decomposes trust-region maintenance into global and local components, providing unified multi-scale stability control for LLM policy optimization. To enhance architectural sparsity, the model comprises 48B total parameters while activating only 2.7B parameters per forward pass, achieving a substantially higher sparsity ratio than contemporary industry leading models of comparable scale. To further improve inference throughput, we adopt a joint training-inference co-design that incorporates dense Multi-Token Prediction (MTP) and Quantization-Aware Training (QAT). We release the checkpoints for both JoyAI-LLM-48B-A3B Base and its post-trained variants on Hugging Face to support the open-source community.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Instruction Following | IFEval | IFEval Accuracy82.8 | 625 | |
| General Knowledge | MMLU | MMLU General Knowledge Accuracy89.1 | 234 | |
| Math | GSM8K | Accuracy0.887 | 206 | |
| Math | MATH 500 | Accuracy98.2 | 86 | |
| Mathematics | MATH | MATH Accuracy78.1 | 85 | |
| Code | HumanEval | HumanEval Accuracy94.5 | 79 | |
| Instruction Following | AlignBench | -- | 60 | |
| Long-context Understanding | RULER | Score95.7 | 50 | |
| General Knowledge Assessment | C-Eval | Accuracy88.7 | 46 | |
| Software Engineering | SWE-bench Verified | Accuracy62.6 | 33 |