Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Subspace Control: Turning Constrained Model Steering into Controllable Spectral Optimization

About

Foundation models, such as large language models (LLMs), are powerful but often require customization before deployment to satisfy practical constraints such as safety, privacy, and task-specific requirements, leading to "constrained" optimization problems for model steering and adaptation. However, solving such problems remains largely underexplored and is particularly challenging due to interference between the primary objective and constraint objectives during optimization. In this paper, we propose a subspace control framework for constrained model training. Specifically, (i) we first analyze, from a model merging perspective, how spectral cross-task interference arises and show that it can be resolved via a one-shot solution that orthogonalizes the merged subspace; (ii) we establish a connection between this solution and gradient orthogonalization in the spectral optimizer Muon; and (iii) building on these insights, we introduce SIFT (spectral interference-free training), which leverages a localization scheme to selectively intervene during optimization, enabling controllable updates that mitigate objective-constraint conflicts. We evaluate SIFT across four representative applications: (a) machine unlearning, (b) safety alignment, (c) text-to-speech adaptation, and (d) hallucination mitigation. Compared to both control-based and control-free baselines, SIFT consistently achieves substantial and robust performance improvements across all tasks. Code is available at https://github.com/OPTML-Group/SIFT.

Yancheng Huang, Changsheng Wang, Chongyu Fan, Yicheng Lang, Bingqi Shang, Yang Zhang, Mingyi Hong, Qing Qu, Alvaro Velasquez, Sijia Liu• 2026

Related benchmarks

TaskDatasetResultRank
Instruction FollowingIFEval
IFEval Accuracy32.7
625
Multi-task Language UnderstandingMMLU
Accuracy56.8
321
Question AnsweringTruthfulQA
Accuracy38.6
152
Natural Language InferenceMNLI--
80
Natural Language InferenceQNLI
Accuracy68.2
61
Safety AlignmentWildJailbreak
Safe@156
24
Language ModelingMMLU
MMLU Final Performance46.2
23
Question AnsweringTruthfulQA
TruthfulQA29
22
Safety AlignmentStrongREJECT--
18
Audio to AudioESNLI (test)
Accuracy77.1
6
Showing 10 of 28 rows

Other info

Follow for update