Split learning (SL) offloads main computing tasks from multiple resource-constrained user equippments (UEs) to the base station (BS), while preserving local data privacy. However, its computation and communication processes remain sequential, resulting in limited system efficiency. To overcome this limitation, this paper applies pipeline parallelism (PP) of distributed training to SL in wireless networks, proposing the so-called communication-computation pipeline parallel split learning (C$^2$P$^2$SL). By considering the communicating and computing processes of UEs and BS as an overall pipeline, C$^2$P$^2$SL achieves pipeline parallelization among different micro-batches which are split from each batch of data samples. The overlap of communication and computation in this way significantly reduces the total training time. Given that training efficiency is affected by position of cutting layer and heterogeneity of the UEs, we formulate a joint optimization problem of task split and resource allocation, and design a solution based on alternating optimization. Experimental results demonstrate that C$^2$P$^2$SL significantly reduces system training time by over 38\% while maintaining convergence accuracy under different communication conditions.
翻译:分割学习(SL)将主要计算任务从多个资源受限的用户设备(UE)卸载至基站(BS),同时保护本地数据隐私。然而,其计算与通信过程仍为串行,导致系统效率受限。为克服此限制,本文将分布式训练的流水线并行(PP)应用于无线网络中的分割学习,提出了通信-计算流水线并行分割学习(C$^2$P$^2$SL)。通过将UE与BS的通信和计算过程视为整体流水线,C$^2$P$^2$SL实现了从每批数据样本划分出的不同微批次之间的流水线并行化。这种通信与计算的重叠显著减少了总体训练时间。考虑到训练效率受分割层位置和UE异构性的影响,我们构建了任务分割与资源分配的联合优化问题,并设计了一种基于交替优化的解决方案。实验结果表明,在不同通信条件下,C$^2$P$^2$SL在保持收敛精度的同时,将系统训练时间显著降低了超过38%。