Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Model Merging in the Essential Subspace

About

Model merging aims to integrate multiple task-specific fine-tuned models derived from a shared pre-trained checkpoint into a single multi-task model without additional training. Despite extensive research, task interference remains a major obstacle that often undermines the performance of merged models. In this paper, we propose ESM (Essential Subspace Merging) , a robust framework for effective model merging. We begin by performing Principal Component Analysis (PCA) on feature shifts induced by parameter updates. The resulting principal directions span an essential subspace that dominantly influences feature representations. Each task's parameter update matrix is projected onto its respective essential subspace for low-rank decomposition before merging. This methodology mitigates inter-task interference while preserving core task-specific functionality. Furthermore, we introduce a multi-level polarized scaling strategy that amplifies parameters containing critical knowledge and suppresses redundant ones, preventing essential knowledge from being overwhelmed during fusion. Extensive experiments across multiple task sets and model scales demonstrate that our method achieves state-of-the-art performance in multi-task model merging.

Longhua Li, Lei Qi, Qi Tian, Xin Geng• 2026

Related benchmarks

TaskDatasetResultRank
Image ClassificationVision Multi-task Suite (SUN397, Cars, RESISC45, EuroSAT, SVHN, GTSRB, MNIST, DTD)
Average Accuracy94.8
72
Image Classification14 Tasks Merge
Average Accuracy91.3
51
Multi-task image classification20-task Model Merging Benchmark (14-task + EMNIST, CIFAR10, Food101, FashionMNIST, RenderedSST2, KMNIST)
Avg Absolute Accuracy90.4
30
Image Classification8-task benchmark
ID Score94.8
6
Showing 4 of 4 rows

Other info

Follow for update