The increasingly deeper neural networks hinder the democratization of privacy-enhancing distributed learning, such as federated learning (FL), to resource-constrained devices. To overcome this challenge, in this paper, we advocate the integration of edge computing paradigm and parallel split learning (PSL), allowing multiple client devices to offload substantial training workloads to an edge server via layer-wise model split. By observing that existing PSL schemes incur excessive training latency and large volume of data transmissions, we propose an innovative PSL framework, namely, efficient parallel split learning (EPSL), to accelerate model training. To be specific, EPSL parallelizes client-side model training and reduces the dimension of local gradients for back propagation (BP) via last-layer gradient aggregation, leading to a significant reduction in server-side training and communication latency. Moreover, by considering the heterogeneous channel conditions and computing capabilities at client devices, we jointly optimize subchannel allocation, power control, and cut layer selection to minimize the per-round latency. Simulation results show that the proposed EPSL framework significantly decreases the training latency needed to achieve a target accuracy compared with the state-of-the-art benchmarks, and the tailored resource management and layer split strategy can considerably reduce latency than the counterpart without optimization.
翻译:不断深化的神经网络阻碍了分布式学习的民主化,例如联邦学习(FL),应用于资源受限的设备。为了克服这一挑战,本文主张整合边缘计算范例和并行分裂学习(PSL),允许多个客户端设备通过分层模型分割将大量的训练工作负载卸载到边缘服务器上。通过观察到现有的PSL方案会产生过多的训练延迟和大量的数据传输,我们提出了一种创新的PSL框架,即高效的并行分裂学习(EPSL),以加速模型训练。具体来说,EPSL并行化了客户端模型训练,并通过最后一层梯度聚合来降低局部梯度的维数,从而大幅减少了服务器端的训练和通信延迟。此外,通过考虑客户端设备的异构信道条件和计算能力,我们联合优化子信道分配、功率控制和切割层数的选择,以最小化每轮的延迟。仿真结果显示,与最先进的基准相比,所提出的EPSL框架显著降低了实现目标精度所需的训练延迟,而定制的资源管理和层分割策略可以比未优化的对应物体显着降低延迟。