Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Solar Open Technical Report

About

We introduce Solar Open, a 102B-parameter bilingual Mixture-of-Experts language model for underserved languages. Solar Open demonstrates a systematic methodology for building competitive LLMs by addressing three interconnected challenges. First, to train effectively despite data scarcity for underserved languages, we synthesize 4.5T tokens of high-quality, domain-specific, and RL-oriented data. Second, we coordinate this data through a progressive curriculum jointly optimizing composition, quality thresholds, and domain coverage across 20 trillion tokens. Third, to enable reasoning capabilities through scalable RL, we apply our proposed framework SnapPO for efficient optimization. Across benchmarks in English and Korean, Solar Open achieves competitive performance, demonstrating the effectiveness of this methodology for underserved language AI development.

Sungrae Park, Sanghoon Kim, Jungho Cho, Gyoungjin Gim, Dawoon Jung, Mikyoung Cha, Eunhae Choo, Taekgyu Hong, Minbyul Jeong, SeHwan Joo, Minsoo Khang, Eunwon Kim, Minjeong Kim, Sujeong Kim, Yunsu Kim, Hyeonju Lee, Seunghyun Lee, Sukyung Lee, Siyoung Park, Gyungin Shin, Inseo Song, Wonho Song, Seonghoon Yang, Seungyoun Yi, Sanghoon Yoon, Jeonghyun Ko, Seyoung Song, Keunwoo Choi, Hwalsuk Lee, Sunghun Kim, Du-Seong Chang, Kyunghyun Cho, Junsuk Choe, Hwaran Lee, Jae-Gil Lee, KyungTae Lim, Alice Oh• 2026

Related benchmarks

TaskDatasetResultRank
Instruction FollowingIFEval--
292
Language UnderstandingMMLU
MMLU Score88.2
45
WritingWritingBench
Score7.51
20
Mathematical Problem SolvingAIME 2025
Score84.3
13
Agentic TaskTau2-Telecom
Accuracy55.6
13
Language UnderstandingMMLU-Pro
MMLU-Pro Score80.4
10
Agentic TasksTau2-Airline
Score52.4
10
Agentic TasksTau2 Retail
Score59.3
10
Chat PreferenceArena-Hard v2
Score79.9
8
Mathematical Problem SolvingHMMT Nov 2025
Score80
8
Showing 10 of 28 rows

Other info

Follow for update