Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

FedProc: Prototypical Contrastive Federated Learning on Non-IID data

About

Federated learning allows multiple clients to collaborate to train high-performance deep learning models while keeping the training data locally. However, when the local data of all clients are not independent and identically distributed (i.e., non-IID), it is challenging to implement this form of efficient collaborative learning. Although significant efforts have been dedicated to addressing this challenge, the effect on the image classification task is still not satisfactory. In this paper, we propose FedProc: prototypical contrastive federated learning, which is a simple and effective federated learning framework. The key idea is to utilize the prototypes as global knowledge to correct the local training of each client. We design a local network architecture and a global prototypical contrastive loss to regulate the training of local models, which makes local objectives consistent with the global optima. Eventually, the converged global model obtains a good performance on non-IID data. Experimental results show that, compared to state-of-the-art federated learning methods, FedProc improves the accuracy by $1.6\%\sim7.9\%$ with acceptable computation cost.

Xutong Mu, Yulong Shen, Ke Cheng, Xueli Geng, Jiaxuan Fu, Tao Zhang, Zhiwei Zhang• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationISIC 2019 (test)--
43
Medical Image ClassificationICH (test)
Balanced Accuracy73.29
11
Medical Image ClassificationReal (last five rounds average)
BACC38.83
11
Showing 3 of 3 rows

Other info

Follow for update