Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Server-Side Local Gradient Averaging and Learning Rate Acceleration for Scalable Split Learning

About

In recent years, there have been great advances in the field of decentralized learning with private data. Federated learning (FL) and split learning (SL) are two spearheads possessing their pros and cons, and are suited for many user clients and large models, respectively. To enjoy both benefits, hybrid approaches such as SplitFed have emerged of late, yet their fundamentals have still been illusive. In this work, we first identify the fundamental bottlenecks of SL, and thereby propose a scalable SL framework, coined SGLR. The server under SGLR broadcasts a common gradient averaged at the split-layer, emulating FL without any additional communication across clients as opposed to SplitFed. Meanwhile, SGLR splits the learning rate into its server-side and client-side rates, and separately adjusts them to support many clients in parallel. Simulation results corroborate that SGLR achieves higher accuracy than other baseline SL methods including SplitFed, which is even on par with FL consuming higher energy and communication costs. As a secondary result, we observe greater reduction in leakage of sensitive information via mutual information using SLGR over the baselines.

Shraman Pal, Mansi Uniyal, Jihong Park, Praneeth Vepakomma, Ramesh Raskar, Mehdi Bennis, Moongu Jeon, Jinho Choi• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-10 IID
Accuracy84.31
166
Image ClassificationCIFAR-100 non-IID (test)
Test Accuracy (Avg Best)41.61
113
Image ClassificationCIFAR-10 non-IID (test)
Average Test Accuracy25.42
14
Image ClassificationCIFAR-10 Mild Non-IID (C=5, α=3.0) (test)
Top-1 Accuracy70.97
13
Image ClassificationCIFAR-10 Severe Non-IID (C=2, alpha=3.0) (test)
Top-1 Accuracy67.73
12
Showing 5 of 5 rows

Other info

Follow for update