Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ProFit: Leveraging High-Value Signals in SFT via Probability-Guided Token Selection

About

Supervised fine-tuning (SFT) is a fundamental post-training strategy to align Large Language Models (LLMs) with human intent. However, traditional SFT often ignores the one-to-many nature of language by forcing alignment with a single reference answer, leading to the model overfitting to non-core expressions. Although our empirical analysis suggests that introducing multiple reference answers can mitigate this issue, the prohibitive data and computational costs necessitate a strategic shift: prioritizing the mitigation of single-reference overfitting over the costly pursuit of answer diversity. To achieve this, we reveal the intrinsic connection between token probability and semantic importance: high-probability tokens carry the core logical framework, while low-probability tokens are mostly replaceable expressions. Based on this insight, we propose ProFit, which selectively masks low-probability tokens to prevent surface-level overfitting. Extensive experiments confirm that ProFit consistently outperforms traditional SFT baselines on general reasoning and mathematical benchmarks.

Tao Liu, Taiqiang Wu, Runming Yang, Shaoning Sun, Junjie Wang, Yujiu Yang• 2026

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy (GSM8K)89.62
358
Instruction FollowingIFEval
Accuracy (0-100)58.02
292
Mathematical ReasoningMATH 500
Accuracy82.85
119
Scientific Question AnsweringGPQA Diamond
Accuracy46.53
64
Multi-task performance evaluationGPQA-Diamond, GSM8K, MATH-500, AIME’24, and IFEval Aggregate
Avg Score58.72
25
Showing 5 of 5 rows

Other info

GitHub

Follow for update