Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Model Fusion via Optimal Transport

About

Combining different models is a widely used paradigm in machine learning applications. While the most common approach is to form an ensemble of models and average their individual predictions, this approach is often rendered infeasible by given resource constraints in terms of memory and computation, which grow linearly with the number of models. We present a layer-wise model fusion algorithm for neural networks that utilizes optimal transport to (soft-) align neurons across the models before averaging their associated parameters. We show that this can successfully yield "one-shot" knowledge transfer (i.e, without requiring any retraining) between neural networks trained on heterogeneous non-i.i.d. data. In both i.i.d. and non-i.i.d. settings , we illustrate that our approach significantly outperforms vanilla averaging, as well as how it can serve as an efficient replacement for the ensemble with moderate fine-tuning, for standard convolutional networks (like VGG11), residual networks (like ResNet18), and multi-layer perceptrons on CIFAR10, CIFAR100, and MNIST. Finally, our approach also provides a principled way to combine the parameters of neural networks with different widths, and we explore its application for model compression. The code is available at the following link, https://github.com/sidak/otfusion.

Sidak Pal Singh, Martin Jaggi• 2019

Related benchmarks

TaskDatasetResultRank
Few-shot classificationMiniImagenet
5-way 5-shot Accuracy27.22
98
Few-shot classificationCUB--
96
Few-shot classificationCIFAR-FS
Accuracy (5-way 1-shot)29.1
58
Node Classification RetentionCora, CiteSeer, Actor, Amazon-Ratings, and Arxiv specialist subsets (held-out)
Retention A92
10
Showing 4 of 4 rows

Other info

Follow for update