While client sampling is a central operation of current state-of-the-art federated learning (FL) approaches, the impact of this procedure on the convergence and speed of FL remains under-investigated. In this work, we provide a general theoretical framework to quantify the impact of a client sampling scheme and of the clients heterogeneity on the federated optimization. First, we provide a unified theoretical ground for previously reported sampling schemes experimental results on the relationship between FL convergence and the variance of the aggregation weights. Second, we prove for the first time that the quality of FL convergence is also impacted by the resulting covariance between aggregation weights. Our theory is general, and is here applied to Multinomial Distribution (MD) and Uniform sampling, two default unbiased client sampling schemes of FL, and demonstrated through a series of experiments in non-iid and unbalanced scenarios. Our results suggest that MD sampling should be used as default sampling scheme, due to the resilience to the changes in data ratio during the learning process, while Uniform sampling is superior only in the special case when clients have the same amount of data.
翻译:虽然客户抽样是目前最先进的联合学习(FL)方法的核心操作,但这一程序对FL的趋同和速度的影响仍然调查不足,在这项工作中,我们提供了一个一般理论框架,以量化客户抽样办法和客户对联邦优化的异质性的影响。第一,我们为以前报告的抽样办法提供了一个统一的理论基础,试验结果涉及FL趋同和总重差异之间的关系。第二,我们第一次证明FL趋同的质量也受到总重之间由此产生的差异的影响。我们的理论是一般性的,在此适用于多数字分布和统一抽样,两种默认的FL不偏重客户抽样办法,并通过一系列非二次和不平衡情况试验加以证明。我们的结果表明,由于学习过程中对数据比率变化的适应力,MD抽样应用作默认的抽样办法,而统一抽样只有在客户拥有相同数量数据的特殊情况下才具有优势。