Federated learning (FL) has emerged as a promising distributed learning paradigm for training deep neural networks (DNNs) at the wireless edge, but its performance can be severely hindered by unreliable wireless transmission and inherent data heterogeneity among clients. Existing solutions primarily address these challenges by incorporating wireless resource optimization strategies, often focusing on uplink resource allocation across clients under the assumption of homogeneous client-server network standards. However, these approaches overlooked the fact that mobile clients may connect to the server via diverse network standards (e.g., 4G, 5G, Wi-Fi) with customized configurations, limiting the flexibility of server-side modifications and restricting applicability in real-world commercial networks. This paper presents a novel theoretical analysis about how transmission failures in unreliable networks distort the effective label distributions of local samples, causing deviations from the global data distribution and introducing convergence bias in FL. Our analysis reveals that a carefully designed client selection strategy can mitigate biases induced by network unreliability and data heterogeneity. Motivated by this insight, we propose FedCote, a client selection approach that optimizes client selection probabilities without relying on wireless resource scheduling. Experimental results demonstrate the robustness of FedCote in DNN-based classification tasks under unreliable networks with frequent transmission failures.
翻译:联邦学习(FL)已成为一种在无线边缘训练深度神经网络(DNN)的有前景的分布式学习范式,但其性能可能因不可靠的无线传输以及客户端间固有的数据异构性而受到严重阻碍。现有解决方案主要通过结合无线资源优化策略来应对这些挑战,通常侧重于在假设客户端-服务器网络标准同质化的前提下,进行跨客户端的上行链路资源分配。然而,这些方法忽略了一个事实,即移动客户端可能通过具有定制配置的不同网络标准(例如,4G、5G、Wi-Fi)连接到服务器,这限制了服务器端修改的灵活性,并制约了其在现实世界商业网络中的适用性。本文提出了一种新颖的理论分析,探讨了不可靠网络中的传输失败如何扭曲本地样本的有效标签分布,导致其偏离全局数据分布,并在联邦学习中引入收敛偏差。我们的分析表明,精心设计的客户端选择策略可以减轻由网络不可靠性和数据异构性引起的偏差。受此启发,我们提出了FedCote,这是一种无需依赖无线资源调度即可优化客户端选择概率的客户端选择方法。实验结果表明,在存在频繁传输失败的不可靠网络中,FedCote在基于DNN的分类任务中表现出鲁棒性。