Federated Learning (FL) has recently received a lot of attention for large-scale privacy-preserving machine learning. However, high communication overheads due to frequent gradient transmissions decelerate FL. To mitigate the communication overheads, two main techniques have been studied: (i) local update of weights characterizing the trade-off between communication and computation and (ii) gradient compression characterizing the trade-off between communication and precision. To the best of our knowledge, studying and balancing those two trade-offs jointly and dynamically while considering their impacts on convergence has remained unresolved even though it promises significantly faster FL. In this paper, we first formulate our problem to minimize learning error with respect to two variables: local update coefficients and sparsity budgets of gradient compression who characterize trade-offs between communication and computation/precision, respectively. We then derive an upper bound of the learning error in a given wall-clock time considering the interdependency between the two variables. Based on this theoretical analysis, we propose an enhanced FL scheme, namely Fast FL (FFL), that jointly and dynamically adjusts the two variables to minimize the learning error. We demonstrate that FFL consistently achieves higher accuracies faster than similar schemes existing in the literature.
翻译:最近,联邦学习联合会(FL)在大规模保护隐私的机器学习方面得到了很多关注,然而,由于频繁的梯度传输而导致的高通信管理费用由于频繁的梯度传播而减速FL。为了减轻通信管理费用,已经研究了两个主要技术:(一) 当地更新通信与计算之间权衡权数的权重,以及(二) 计算通信与精确之间的权衡权重的梯度压缩。为了最佳的知识,在考虑这两个取舍对趋同的影响的同时,共同和动态地研究并平衡这两个取舍,尽管它承诺大大加快FL。 在本文件中,我们首先提出问题,尽量减少两个变量的学习错误:当地更新系数和梯度压缩预算,分别说明通信与计算/精度之间的利差。然后,考虑到两个变量之间的相互依存关系,我们从学习错误的上限上拉开。根据这一理论分析,我们提出一个强化的FL计划,即快速FL(FFL),即联合和动态地调整两个变量,以比现有文献的更快的速度来尽量减少学习错误。