The increasingly deeper neural networks hinder the democratization of privacy-enhancing distributed learning, such as federated learning (FL), to resource-constrained devices. To overcome this challenge, in this paper, we advocate the integration of edge computing paradigm and parallel split learning (PSL), allowing multiple client devices to offload substantial training workloads to an edge server via layer-wise model split. By observing that existing PSL schemes incur excessive training latency and large volume of data transmissions, we propose an innovative PSL framework, namely, efficient parallel split learning (EPSL), to accelerate model training. To be specific, EPSL parallelizes client-side model training and reduces the dimension of local gradients for back propagation (BP) via last-layer gradient aggregation, leading to a significant reduction in server-side training and communication latency. Moreover, by considering the heterogeneous channel conditions and computing capabilities at client devices, we jointly optimize subchannel allocation, power control, and cut layer selection to minimize the per-round latency. Simulation results show that the proposed EPSL framework significantly decreases the training latency needed to achieve a target accuracy compared with the state-of-the-art benchmarks, and the tailored resource management and layer split strategy can considerably reduce latency than the counterpart without optimization.
翻译:随着神经网络越来越深,阻碍了隐私增强分布式学习(例如联邦学习)在资源限制的设备上进行普及。为了克服这个挑战,在本文中,我们提倡整合边缘计算范例和并行分裂学习(PSL)来破解这一难题,使多个客户端设备通过分层模型分裂向边缘服务器卸载大量的训练工作负载。通过观察现有的 PSL 方案存在过多的训练延迟和大量的数据传输,我们提出了一种创新的 PSL 框架 EPSL,来加速模型训练。具体而言,EPSL 并行化了客户端模型训练,通过最后一层梯度聚合,降低了反向传播(BP)的本地梯度的维数,从而显著降低了服务器端训练和通信延迟。此外,通过考虑客户端设备的异构信道条件和计算能力,我们联合优化子信道分配、功率控制和分层选择,以最小化每轮延迟。仿真结果显示,与现有的基准相比,所提出的 EPSL 框架显著缩短了达到目标精度所需的训练延迟,所制定的资源管理和分层分裂策略比没有进行优化的对应方案大幅降低了延迟。