Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Local Superior Soups: A Catalyst for Model Merging in Cross-Silo Federated Learning

About

Federated learning (FL) is a learning paradigm that enables collaborative training of models using decentralized data. Recently, the utilization of pre-trained weight initialization in FL has been demonstrated to effectively improve model performance. However, the evolving complexity of current pre-trained models, characterized by a substantial increase in parameters, markedly intensifies the challenges associated with communication rounds required for their adaptation to FL. To address these communication cost issues and increase the performance of pre-trained model adaptation in FL, we propose an innovative model interpolation-based local training technique called ``Local Superior Soups.'' Our method enhances local training across different clients, encouraging the exploration of a connected low-loss basin within a few communication rounds through regularized model interpolation. This approach acts as a catalyst for the seamless adaptation of pre-trained models in in FL. We demonstrated its effectiveness and efficiency across diverse widely-used FL datasets. Our code is available at \href{https://github.com/ubc-tea/Local-Superior-Soups}{https://github.com/ubc-tea/Local-Superior-Soups}.

Minghui Chen, Meirui Jiang, Xin Zhang, Qi Dou, Zehua Wang, Xiaoxiao Li• 2024

Related benchmarks

TaskDatasetResultRank
Image ClassificationDomainNet (test)
Average Accuracy41.35
209
Digit ClassificationDigit-Five (test)
Average Accuracy92.97
60
Image ClassificationFMNIST label shift (test)
Top-1 Accuracy72.66
12
Image ClassificationCIFAR-10 label shift (test)
Top-1 Accuracy65.96
12
Image ClassificationDigit-5 feature shift (test)
Accuracy (R=1)72.86
12
Image ClassificationDomainNet feature shift (test)
Accuracy (R=1)27.86
12
Showing 6 of 6 rows

Other info

Code

Follow for update