Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Taming Cross-Domain Representation Variance in Federated Prototype Learning with Heterogeneous Data Domains

About

Federated learning (FL) allows collaborative machine learning training without sharing private data. While most FL methods assume identical data domains across clients, real-world scenarios often involve heterogeneous data domains. Federated Prototype Learning (FedPL) addresses this issue, using mean feature vectors as prototypes to enhance model generalization. However, existing FedPL methods create the same number of prototypes for each client, leading to cross-domain performance gaps and disparities for clients with varied data distributions. To mitigate cross-domain feature representation variance, we introduce FedPLVM, which establishes variance-aware dual-level prototypes clustering and employs a novel $\alpha$-sparsity prototype loss. The dual-level prototypes clustering strategy creates local clustered prototypes based on private data features, then performs global prototypes clustering to reduce communication complexity and preserve local data privacy. The $\alpha$-sparsity prototype loss aligns samples from underrepresented domains, enhancing intra-class similarity and reducing inter-class similarity. Evaluations on Digit-5, Office-10, and DomainNet datasets demonstrate our method's superiority over existing approaches.

Lei Wang, Jieming Bian, Letian Zhang, Chen Chen, Jie Xu• 2024

Related benchmarks

TaskDatasetResultRank
Image ClassificationDomainNet (test)
Average Accuracy42.55
209
Digit ClassificationDigit-Five (test)
Average Accuracy69.25
60
Image ClassificationOffice-Caltech-10 (test)--
27
Showing 3 of 3 rows

Other info

Follow for update