Accelerating Split Federated Learning over Wireless Communication Networks
About
The development of artificial intelligence (AI) provides opportunities for the promotion of deep neural network (DNN)-based applications. However, the large amount of parameters and computational complexity of DNN makes it difficult to deploy it on edge devices which are resource-constrained. An efficient method to address this challenge is model partition/splitting, in which DNN is divided into two parts which are deployed on device and server respectively for co-training or co-inference. In this paper, we consider a split federated learning (SFL) framework that combines the parallel model training mechanism of federated learning (FL) and the model splitting structure of split learning (SL). We consider a practical scenario of heterogeneous devices with individual split points of DNN. We formulate a joint problem of split point selection and bandwidth allocation to minimize the system latency. By using alternating optimization, we decompose the problem into two sub-problems and solve them optimally. Experiment results demonstrate the superiority of our work in latency reduction and accuracy improvement.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-100 (test) | Top-1 Acc64.08 | 287 | |
| Image Classification | CIFAR-10 (test) | Accuracy90.94 | 19 | |
| Model Training | CIFAR-10 (non-IID) | Training Delay (minutes)3.27e+3 | 16 | |
| Model Training | CIFAR-10 IID | Training Delay (minutes)2.87e+3 | 16 | |
| Model Training | CIFAR-100 IID | Training Delay (minutes)2.79e+3 | 16 | |
| Model Training | CIFAR-100 (non-IID) | Training Delay (minutes)3.33e+3 | 16 |