Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Orthogonal Model Merging

About

Merging finetuned Large Language Models (LLMs) has become increasingly important for integrating diverse capabilities into a single unified model. However, prevailing model merging methods rely on linear arithmetic in Euclidean space, which often destroys the intrinsic geometric properties of pretrained weights, such as hyperspherical energy. To address this, we propose Orthogonal Model Merging (OrthoMerge), a method that performs merging operations on the Riemannian manifold formed by the orthogonal group to preserve the geometric structure of the model's weights. By mapping task-specific orthogonal matrices learned by Orthogonal Finetuning (OFT) to the Lie algebra, OrthoMerge enables a principled yet efficient integration that takes into account both the direction and intensity of adaptations. In addition to directly leveraging orthogonal matrices obtained by OFT, we further extend this approach to general models finetuned with non-OFT methods (i.e., low-rank finetuning, full finetuning) via an Orthogonal-Residual Decoupling strategy. This technique extracts the orthogonal components of expert models by solving the orthogonal Procrustes problem, which are then merged on the manifold of the orthogonal group, while the remaining linear residuals are processed through standard additive merging. Extensive empirical results demonstrate the effectiveness of OrthoMerge in mitigating catastrophic forgetting and maintaining model performance across diverse tasks.

Sihan Yang, Kexuan Shi, Weiyang Liu• 2026

Related benchmarks

TaskDatasetResultRank
Science Question AnsweringScienceQA
Accuracy32.01
229
Code GenerationHumanEval+--
189
Social Interaction Question AnsweringSIQA
Accuracy55.17
85
General ReasoningAGIEval
Exact Match38.84
33
Social Intelligence Question AnsweringSocial-IQA
Score52.66
12
Multi-task Language ModelingMergeBench
Instruction Score25.32
11
Vision-Language Multi-task PerformanceMergeBench (Vision-Language tasks: MMSI-Bench, EmbSpatial, MMMU_Med, PathVQA, OCRBench, CharXiv)
MMSI-Bench32.5
11
General Language UnderstandingMMLU, AGIEval
MMLU Score0.5592
10
Multilingual Question Answeringm-ARC
Accuracy44.75
10
Instruction Following and Multimodal ReasoningGeneralist VLM Benchmarks IFEval, MMBench
IFEval Score61
10
Showing 10 of 10 rows

Other info

Follow for update