Federated Learning (FL) enables mobile edge devices, functioning as clients, to collaboratively train a decentralized model while ensuring local data privacy. However, the efficiency of FL in wireless networks is limited not only by constraints on communication and computational resources but also by significant data heterogeneity among clients, particularly in large-scale networks. This paper first presents a theoretical analysis of the impact of client data heterogeneity on global model generalization error, which can result in repeated training cycles, increased energy consumption, and prolonged latency. Based on the theoretical insights, an optimization problem is formulated to jointly minimize learning latency and energy consumption while constraining generalization error. A joint client selection and resource allocation (CSRA) approach is then proposed, employing a series of convex optimization and relaxation techniques. Extensive simulation results demonstrate that the proposed CSRA scheme yields higher test accuracy, reduced learning latency, and lower energy consumption compared to baseline methods that do not account for data heterogeneity.
翻译:联邦学习(FL)使得作为客户端的移动边缘设备能够在保证本地数据隐私的前提下,协作训练一个去中心化模型。然而,无线网络中联邦学习的效率不仅受到通信和计算资源的限制,还受到客户端之间显著的数据异构性(尤其是在大规模网络中)的制约。本文首先从理论上分析了客户端数据异构性对全局模型泛化误差的影响,这种影响可能导致训练周期重复、能耗增加以及时延延长。基于理论分析,我们构建了一个优化问题,旨在约束泛化误差的同时联合最小化学习时延与能耗。随后,提出了一种联合客户端选择与资源分配(CSRA)方法,该方法采用了一系列凸优化与松弛技术。大量仿真结果表明,与未考虑数据异构性的基准方法相比,所提出的CSRA方案能够实现更高的测试精度、更低的学习时延以及更低的能耗。