Federated averaging (FedAvg) is a popular federated learning (FL) technique that updates the global model by averaging local models and then transmits the updated global model to devices for their local model update. One main limitation of FedAvg is that the average-based global model is not necessarily better than local models in the early stage of the training process so that FedAvg might diverge in realistic scenarios, especially when the data is non-identically distributed across devices and the number of data samples varies significantly from device to device. In this paper, we propose a new FL technique based on simulated annealing. The key idea of the proposed technique, henceforth referred to as \textit{simulated annealing-based FL} (SAFL), is to allow a device to choose its local model when the global model is immature. Specifically, by exploiting the simulated annealing strategy, we make each device choose its local model with high probability in early iterations when the global model is immature. From extensive numerical experiments using various benchmark datasets, we demonstrate that SAFL outperforms the conventional FedAvg technique in terms of the convergence speed and the classification accuracy.
翻译:联邦平均(FedAvg)是一种广受欢迎的联邦学习(FL)技术,它通过平均当地模型来更新全球模型,然后将更新的全球模型传送到本地模型更新的装置。FedAvg的一个主要限制是,在培训进程的早期阶段,基于平均的全球模型不一定比基于培训过程早期阶段的当地模型好,这样FedAvg在现实的假设中可能会有差异,特别是当数据在各种设备之间不明显地分布,数据样本数量在设备之间差异很大时。在本文中,我们提出一个新的基于模拟肛交的新的FL技术。拟议技术的主要构想,即今后称为\ textit{模拟以射线为基础的FL}(SAFL),是让一个设备在全球模型不成熟时选择其本地模型。具体地说,我们利用模拟的反射战略,使每个装置选择其本地模型的概率很高,当全球模型不成熟时,我们从使用各种基准数据集进行广泛的数字实验中,我们证明SAFL在常规的分类中超过了常规的精确性。