Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Condor: Enhance LLM Alignment with Knowledge-Driven Data Synthesis and Refinement

About

The quality of Supervised Fine-Tuning (SFT) data plays a critical role in enhancing the conversational capabilities of Large Language Models (LLMs). However, as LLMs become more advanced, the availability of high-quality human-annotated SFT data has become a significant bottleneck, necessitating a greater reliance on synthetic training data. In this work, we introduce Condor, a novel two-stage synthetic data generation framework that incorporates World Knowledge Tree and Self-Reflection Refinement to produce high-quality SFT data at scale. Our experimental results demonstrate that a base model fine-tuned on only 20K Condor-generated samples achieves superior performance compared to counterparts. The additional refinement stage in Condor further enables iterative self-improvement for LLMs at various scales (up to 72B), validating the effectiveness of our approach. Furthermore, our investigation into the scaling for synthetic data in post-training reveals substantial unexplored potential for performance improvements, opening promising avenues for future research.

Maosong Cao, Taolin Zhang, Mo Li, Chuyu Zhang, Yunxin Liu, Haodong Duan, Songyang Zhang, Kai Chen• 2025

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval
Pass@151.83
850
Mathematical ReasoningGSM8K
Accuracy61.49
351
Mathematical Problem SolvingMATH
Accuracy48.6
166
Showing 3 of 3 rows

Other info

Follow for update