Cognitively Layered Data Synthesis for Domain Adaptation of LLMs to Space Situational Awareness
About
Large language models (LLMs) demonstrate exceptional performance on general-purpose tasks. however, transferring them to complex engineering domains such as space situational awareness (SSA) remains challenging owing to insufficient structural alignment with mission chains, the absence of higher-order cognitive supervision, and poor correspondence between data quality criteria and engineering specifications. The core bottleneck is the construction of high-quality supervised fine-tuning (SFT) datasets. To this end, we propose BD-FDG (Bloom's Taxonomy-based Domain-specific Fine-tuning Data Generation), a framework that addresses incomplete knowledge coverage, shallow cognitive depth, and limited quality controllability through three mechanisms: structured knowledge organization, cognitively layered question modeling, and automated quality control. The framework uses a knowledge tree to ensure structured corpus coverage, designs a question generation scheme spanning nine categories and six cognitive levels from Remember to Create to produce samples with a continuous difficulty gradient, and applies a multidimensional scoring pipeline to enforce domain rigor and consistency. Using BD-FDG, we construct SSA-SFT, a domain dataset of approximately 230K samples, and fine-tune Qwen3-8B to obtain SSA-LLM-8B. Experiments show that SSA-LLM-8B achieves relative BLEU-1 improvements of 144\% (no-think) and 176\% (think) on the domain test set and a win rate of 82.21\% over the baseline in arena comparisons, while largely preserving general benchmark performance (MMLU-Pro, MATH-500). These results validate SFT data construction driven by cognitive layering as an effective paradigm for complex engineering domains and provide a transferable framework for domain-specific LLM adaptation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Instruction Following | IFEval | IFEval Accuracy79.66 | 625 | |
| Math | MATH 500 | Accuracy94.8 | 86 | |
| Mathematics | AIME25 | Accuracy56.67 | 63 | |
| Math | AIME24 | Accuracy66.67 | 38 | |
| MMLU-Redux | MMLU-Redux | Accuracy87.47 | 17 | |
| Examination | MMLU-Pro | Score0.7315 | 9 | |
| Coding | LiveCodeBench | Accuracy40.11 | 8 | |
| Knowledge | SSA (test) | BLEU-157.23 | 4 | |
| Exam | iQuiz | Accuracy60.83 | 4 | |
| Knowledge | GPQA Diamond | Accuracy (GPQA Knowledge)60.1 | 4 |