Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Merging Models on the Fly Without Retraining: A Sequential Approach to Scalable Continual Model Merging

About

Deep model merging represents an emerging research direction that combines multiple fine-tuned models to harness their specialized capabilities across different tasks and domains. Current model merging techniques focus on merging all available models simultaneously, with weight interpolation-based methods being the predominant approaches. However, these conventional approaches are not well-suited for scenarios where models become available sequentially, and they often suffer from high memory requirements and potential interference between tasks. In this study, we propose a training-free projection-based continual merging method that processes models sequentially through orthogonal projections of weight matrices and adaptive scaling mechanisms. Our method operates by projecting new parameter updates onto subspaces orthogonal to existing merged parameter updates while using an adaptive scaling mechanism to maintain stable parameter distances, enabling efficient sequential integration of task-specific knowledge. Our approach maintains constant memory complexity to the number of models, minimizes interference between tasks through orthogonal projections, and retains the performance of previously merged models through adaptive task vector scaling. Extensive experiments on CLIP-ViT models demonstrate that our method achieves a 5-8% average accuracy improvement while maintaining robust performance in different task orderings.

Anke Tang, Enneng Yang, Li Shen, Yong Luo, Han Hu, Bo Du, Dacheng Tao• 2025

Related benchmarks

TaskDatasetResultRank
Continual LearningLarge Number of Tasks
Average Performance50.5
50
Continual LearningStandard CL Benchmark
Avg Final Acc0.602
50
Continual LearningSuperNI Benchmark
Average Score12
14
Continual LearningLarge Number of Tasks (test)
Backward Transfer (BWT)-3.9
13
Continual LearningSuperNI Standard CL Benchmark (test)
Average Performance64.7
13
Continual LearningSuperNI Large Number of Tasks (test)
Average Performance58.9
13
Showing 6 of 6 rows

Other info

Follow for update