The statistical heterogeneity of the non-independent and identically distributed (non-IID) data in local clients significantly limits the performance of federated learning. Previous attempts like FedProx, SCAFFOLD, MOON, FedNova and FedDyn resort to an optimization perspective, which requires an auxiliary term or re-weights local updates to calibrate the learning bias or the objective inconsistency. However, in addition to previous explorations for improvement in federated averaging, our analysis shows that another critical bottleneck is the poorer optima of client models in more heterogeneous conditions. We thus introduce a data-driven approach called FedSkip to improve the client optima by periodically skipping federated averaging and scattering local models to the cross devices. We provide theoretical analysis of the possible benefit from FedSkip and conduct extensive experiments on a range of datasets to demonstrate that FedSkip achieves much higher accuracy, better aggregation efficiency and competing communication efficiency. Source code is available at: https://github.com/MediaBrain-SJTU/FedSkip.
翻译:本地客户中非独立和同样分布(非IID)的统计数据差异性大大限制了联合学习的绩效。之前的尝试,如FedProx、SCAFFOLD、MOON、FedNova和FedDyn, 都采用了优化观点,这要求当地进行辅助性术语或再加权更新,以校准学习偏差或客观不一致之处。然而,除了以往对改进联邦平均比例的探索外,我们的分析表明,另一个关键瓶颈是客户模型在更多样化的条件下的更差的选用。我们因此采用了一种数据驱动方法,称为FedSkip,通过定期跳过联邦平均和向交叉设备传播当地模型来改进客户的选用。我们从理论上分析了FedSkip可能带来的好处,并在一系列数据集上进行了广泛的实验,以证明FedSkip实现了更高的准确性、更好的汇总效率和竞争性通信效率。源代码见:https://github.com/Medibrain-SJTU/FedSkip。