Federated learning (FL) algorithms usually sample a fraction of clients in each round (partial participation) when the number of participants is large and the server's communication bandwidth is limited. Recent works on the convergence analysis of FL have focused on unbiased client sampling, e.g., sampling uniformly at random, which suffers from slow wall-clock time for convergence due to high degrees of system heterogeneity and statistical heterogeneity. This paper aims to design an adaptive client sampling algorithm that tackles both system and statistical heterogeneity to minimize the wall-clock convergence time. We obtain a new tractable convergence bound for FL algorithms with arbitrary client sampling probabilities. Based on the bound, we analytically establish the relationship between the total learning time and sampling probabilities, which results in a non-convex optimization problem for training time minimization. We design an efficient algorithm for learning the unknown parameters in the convergence bound and develop a low-complexity algorithm to approximately solve the non-convex problem. Experimental results from both hardware prototype and simulation demonstrate that our proposed sampling scheme significantly reduces the convergence time compared to several baseline sampling schemes. Notably, our scheme in hardware prototype spends 73% less time than the uniform sampling baseline for reaching the same target loss.
翻译:联邦学习算法(FL)通常在参与者人数众多且服务器通信带宽有限的情况下对每轮(部分参与)客户进行抽样(部分参与),当参与者人数众多且服务器通信带宽有限时,每轮(部分参与)客户的一小部分(部分参与),最近关于FL趋同分析的工作侧重于不带偏见的客户抽样,例如,统一随机抽样,由于系统差异性和统计差异性高,趋同时间慢于墙上最短的钟点,因为系统差异性和统计差异性较高,结果造成聚集时间的慢点,结果,本文旨在设计适应性客户抽样算法,既处理系统和统计差异性,又尽量减少墙时钟汇合时间。我们获得了FL算法与任意客户抽样概率的新的可移植趋同点。我们根据这一趋同点,分析确定了总学习时间与取样概率之间的关系,从而造成培训时间最小化的非凝聚点优化问题。我们设计了一种有效的算法,用于了解趋同点的未知参数,并开发一种低相兼容性算法,以尽量解决非凝点问题。我们提议的硬件原型和模拟的实验结果表明,我们提议的取样计划大大缩短了接近接近接近率的时间,比73基准取样率的基数率。