In federated learning (FL) problems, client sampling plays a key role in the convergence speed of training algorithm. However, while being an important problem in FL, client sampling is lack of study. In this paper, we propose an online learning with bandit feedback framework to understand the client sampling problem in FL. By adapting an Online Stochastic Mirror Descent algorithm to minimize the variance of gradient estimation, we propose a new adaptive client sampling algorithm. Besides, we use online ensemble method and doubling trick to automatically choose the tuning parameters in the algorithm. Theoretically, we show dynamic regret bound with comparator as the theoretically optimal sampling sequence; we also include the total variation of this sequence in our upper bound, which is a natural measure of the intrinsic difficulty of the problem. To the best of our knowledge, these theoretical contributions are novel to existing literature. Moreover, by implementing both synthetic and real data experiments, we show empirical evidence of the advantages of our proposed algorithms over widely-used uniform sampling and also other online learning based sampling strategies in previous studies. We also examine its robustness to the choice of tuning parameters. Finally, we discuss its possible extension to sampling without replacement and personalized FL objective. While the original goal is to solve client sampling problem, this work has more general applications on stochastic gradient descent and stochastic coordinate descent methods.
翻译:在联合学习(FL)问题中,客户抽样在培训算法的趋同速度方面起着关键作用。然而,虽然客户抽样是FL的一个重要问题,但客户抽样却缺乏研究。在本文中,我们提议用土匪反馈框架进行在线学习,以了解FL的客户抽样问题。通过调整在线斯托切斯镜源算法,以尽量减少梯度估计的差异,我们提出了一种新的适应性客户抽样算法。此外,我们使用在线混合法和双倍技巧,以自动选择算法中的调试参数。理论上,我们表现出与比较国作为理论上最佳抽样序列的动态遗憾;我们还把这种序列的全变换纳入我们的上层,这是衡量问题内在困难的自然尺度。根据我们的最佳知识,这些理论贡献对现有文献来说是新颖的。此外,通过实施合成和真实的数据实验,我们展示了我们提议的算法在以往研究中比广泛使用的统一抽样和其他基于在线学习的抽样战略的优势。我们还检查了它是否稳健地选择了调整参数。最后,我们讨论了将其扩展为原始取样的升级目标,与此同时,我们讨论了其可能扩展为原客户抽样,而没有先变换为原的底目标,而先变为原的客户则则则比较目标。