Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Federated Generalized Expectation-Maximization Algorithm for Mixture Models with an Unknown Number of Components

About

We study the problem of federated clustering when the total number of clusters $K$ across clients is unknown, and the clients have heterogeneous but potentially overlapping cluster sets in their local data. To that end, we develop FedGEM: a federated generalized expectation-maximization algorithm for the training of mixture models with an unknown number of components. Our proposed algorithm relies on each of the clients performing EM steps locally, and constructing an uncertainty set around the maximizer associated with each local component. The central server utilizes the uncertainty sets to learn potential cluster overlaps between clients, and infer the global number of clusters via closed-form computations. We perform a thorough theoretical study of our algorithm, presenting probabilistic convergence guarantees under common assumptions. Subsequently, we study the specific setting of isotropic GMMs, providing tractable, low-complexity computations to be performed by each client during each iteration of the algorithm, as well as rigorously verifying assumptions required for algorithm convergence. We perform various numerical experiments, where we empirically demonstrate that our proposed method achieves comparable performance to centralized EM, and that it outperforms various existing federated clustering methods.

Michael Ibrahim, Nagi Gebraeel, Weijun Xie• 2026

Related benchmarks

TaskDatasetResultRank
Image ClusteringCIFAR-10--
243
ClusteringFMNIST--
31
ClusteringMNIST
ARI0.452
19
ClusteringEMNIST
ARI28.5
17
ClusteringWaveform
Adjusted Rand Index0.335
8
ClusteringWaveform
SS0.271
8
ClusteringMNIST
Silhouette Score0.095
8
ClusteringFrog A
ARI55.2
8
ClusteringFrog B
ARI0.468
8
ClusteringEMNIST
SS0.063
8
Showing 10 of 21 rows

Other info

Follow for update