Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

POME: Post Optimization Model Edit via Muon-style Projection

About

We introduce Post-Optimization Model Edit (POME), a new algorithm that enhances the performance of fine-tuned large language models using only their pretrained and fine-tuned checkpoints, without requiring extra data or further optimization. The core idea is to apply a muon-style projection to $\Delta W$, the difference between the fine-tuned and pretrained weights. This projection uses truncated singular value decomposition (SVD) to equalize the influence of dominant update directions and prune small singular values, which often represent noise. As a simple post-processing step, POME is completely decoupled from the training pipeline. It requires zero modifications and imposes no overhead, making it universally compatible with any optimizer or distributed framework. POME delivers consistent gains, boosting average performance by +2.5\% on GSM8K and +1.0\% on code generation. Its broad applicability -- from 7B foundation models to 72B RLHF-instructed models -- establishes it as a practical, zero-cost enhancement for any fine-tuning pipeline. Code is available at https://github.com/NUS-HPC-AI-Lab/POME.

Yong Liu, Di Fu, Yang Luo, Zirui Zhu, Minhao Cheng, Cho-Jui Hsieh, Yang You• 2025

Related benchmarks

TaskDatasetResultRank
Instruction FollowingIFEval
IFEval Accuracy31.6
625
Multi-task Language UnderstandingMMLU
Accuracy57.3
321
Question AnsweringTruthfulQA
Accuracy38.4
152
Natural Language InferenceMNLI--
80
Natural Language InferenceQNLI
Accuracy68.2
61
Safety AlignmentWildJailbreak
Safe@151.6
24
Language ModelingMMLU
MMLU Final Performance46
23
Question AnsweringTruthfulQA
TruthfulQA29.2
22
Safety AlignmentStrongREJECT--
18
Text-to-TextESNLI (test)
Accuracy (ESNLI Test)80.6
6
Showing 10 of 28 rows

Other info

Follow for update