Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ERIS: Enhancing Privacy and Communication Efficiency in Serverless Federated Learning

About

Scaling federated learning (FL) to billion-parameter models introduces critical trade-offs between communication efficiency, model accuracy, and privacy guarantees. Existing solutions often tackle these challenges in isolation, sacrificing accuracy or relying on costly cryptographic tools. We propose ERIS, a serverless FL framework that balances privacy and accuracy while eliminating the server bottleneck and distributing the communication load. ERIS combines a model partitioning strategy, distributing aggregation across multiple client-side aggregators, with a distributed shifted gradient compression mechanism. We theoretically prove that ERIS (i) converges at the same rate as FedAvg under standard assumptions, and (ii) bounds mutual information leakage inversely with the number of aggregators, enabling strong privacy guarantees with no accuracy degradation. Experiments across image and text tasks, including large language models, confirm that ERIS achieves FedAvg-level accuracy while substantially reducing communication cost and improving robustness to membership inference and reconstruction attacks, without relying on heavy cryptography or noise injection.

Dario Fenoglio, Pasquale Polverino, Jacopo Quizi, Martin Gjoreski, Marc Langheinrich• 2026

Related benchmarks

TaskDatasetResultRank
Text ClassificationIMDB (test)
Accuracy79.07
18
Image ClassificationCIFAR-10 (test)
Accuracy30.16
9
Image ClassificationMNIST (test)
Accuracy85.1
9
Text SummarizationCNN/DailyMail (test)
ROUGE-132.83
9
Image ClassificationCIFAR-10
Distribution Time (s)0.0039
7
SummarizationCNN/DailyMail
Distribution Time (s)4.68
7
Showing 6 of 6 rows

Other info

Follow for update