Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Model-Contrastive Federated Learning

About

Federated learning enables multiple parties to collaboratively train a machine learning model without communicating their local data. A key challenge in federated learning is to handle the heterogeneity of local data distribution across parties. Although many studies have been proposed to address this challenge, we find that they fail to achieve high performance in image datasets with deep learning models. In this paper, we propose MOON: model-contrastive federated learning. MOON is a simple and effective federated learning framework. The key idea of MOON is to utilize the similarity between model representations to correct the local training of individual parties, i.e., conducting contrastive learning in model-level. Our extensive experiments show that MOON significantly outperforms the other state-of-the-art federated learning algorithms on various image classification tasks.

Qinbin Li, Bingsheng He, Dawn Song• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy19.73
3518
Image ClassificationCIFAR-10 (test)
Accuracy54.49
3381
Object Hallucination EvaluationPOPE
Accuracy74.1
1455
Image ClassificationCIFAR-10 (test)--
906
Image ClassificationCIFAR-100--
691
Multimodal EvaluationMME--
658
Multimodal UnderstandingMMBench
Accuracy34.5
637
Image ClassificationCIFAR10 (test)
Accuracy57.62
585
Image ClassificationCIFAR-10
Accuracy69.24
564
Person Re-IdentificationMSMT17
mAP0.048
514
Showing 10 of 211 rows
...

Other info

Code

Follow for update