Deploying foundation models (FMs) on uncrewed aerial vehicles (UAVs) promises broad ``low-altitude economy'' applications. Split federated learning (SFL)-based fine-tuning leverages distributed data while keeping raw data local and reduces client-side burden by partitioning the model between client and server. However, the per-round training latency is dominated by stragglers. Training paradigms featuring parallel gradient transmission (GT) allocate dedicated portions of downlink communication resources to each client. They may leave resources idle and suffer from prolonged GT latency, especially in UAV networks, where the communication latency typically far exceeds the computation latency. To address this, we propose a sequential GT paradigm, where the server dedicates all downlink resources for the current GT. We further propose communication-pipelined SFL (CPSFL), characterized by downlink GT priority scheduling and intra-round asynchronous training. We investigate CPSFL-based LoRA fine-tuning of FMs in UAV networks and formulate an optimization problem to minimize a weighted sum of per-round training latency and worst-case client energy consumption by optimizing the split point selection (SPS) and the computing and communication resource allocation (CCRA) (the uplink bandwidth allocation and the server computing frequency allocation). To solve this problem, we develop an attention-based deep reinforcement learning (DRL) framework, where the base station agent decides the split point and the CCRA in each round by leveraging previous round information, including UAV trajectories. Simulation results show that the proposed DRL-based CPSFL scheme outperforms the parallel GT benchmarks, the ablation variants, the fixed CCRA scheme, while approaching the best fixed-SPS scheme.
翻译:在无人机上部署基础模型有望推动广泛的“低空经济”应用。基于拆分联邦学习的微调方法利用分布式数据,同时保持原始数据本地化,并通过在客户端与服务器之间划分模型来减轻客户端负担。然而,每轮训练延迟主要受制于掉队者。采用并行梯度传输的训练范式为每个客户端分配专用的下行通信资源部分,可能导致资源闲置并遭受梯度传输延迟延长,尤其在无人机网络中,通信延迟通常远超计算延迟。为解决此问题,我们提出一种顺序梯度传输范式,其中服务器将所有下行资源专用于当前梯度传输。我们进一步提出通信流水线拆分联邦学习,其特点为下行梯度传输优先级调度与轮内异步训练。我们研究了无人机网络中基于CPSFL的LoRA基础模型微调,并构建了一个优化问题,通过优化拆分点选择以及计算与通信资源分配(包括上行带宽分配和服务器计算频率分配),以最小化每轮训练延迟与最差客户端能耗的加权和。为解决该问题,我们开发了一个基于注意力的深度强化学习框架,其中基站代理利用包括无人机轨迹在内的前轮信息,在每轮中决定拆分点与资源分配。仿真结果表明,所提出的基于DRL的CPSFL方案在性能上优于并行梯度传输基准方案、消融变体方案、固定资源分配方案,并接近最佳固定拆分点方案。