Federated learning (FL) enables distributed model training from local data collected by users. In distributed systems with constrained resources and potentially high dynamics, e.g., mobile edge networks, the efficiency of FL is an important problem. Existing works have separately considered different configurations to make FL more efficient, such as infrequent transmission of model updates, client subsampling, and compression of update vectors. However, an important open problem is how to jointly apply and tune these control knobs in a single FL algorithm, to achieve the best performance by allowing a high degree of freedom in control decisions. In this paper, we address this problem and propose FlexFL - an FL algorithm with multiple options that can be adjusted flexibly. Our FlexFL algorithm allows both arbitrary rates of local computation at clients and arbitrary amounts of communication between clients and the server, making both the computation and communication resource consumption adjustable. We prove a convergence upper bound of this algorithm. Based on this result, we further propose a stochastic optimization formulation and algorithm to determine the control decisions that (approximately) minimize the convergence bound, while conforming to constraints related to resource consumption. The advantage of our approach is also verified using experiments.
翻译:联邦学习(FL) 能够利用用户收集的地方数据进行分布式示范培训。在资源有限且具有潜在高度动态的分布式系统中,例如移动边缘网络,FL的效率是一个重要问题。现有的工程分别考虑不同的配置,以提高FL效率,例如不经常传输模型更新、客户子抽样和压缩更新矢量。然而,一个重要的未决问题是如何在单一FL算法中联合应用和调控这些控制 knob,以便通过允许高度自由控制决策实现最佳性能。在本文中,我们解决这个问题并提出FlexFL - 一种可灵活调整多种选项的FL算法。我们的FlexFL算法允许本地客户任意计算率和客户与服务器之间任意的通信量,使计算和通信资源消耗量可以调整。我们证明这种算法的高度界限是趋同。我们进一步提议一种随机优化的公式和算法,以确定控制决定(约) 尽量减少约束的趋同,同时符合资源消耗的限制。我们的方法的优势也是通过实验加以核查。