As a promising learning paradigm integrating computation and communication, federated learning (FL) proceeds the local training and the periodic sharing from distributed clients. Due to the non-i.i.d. data distribution on clients, FL model suffers from the gradient diversity, poor performance, bad convergence, etc. In this work, we aim to tackle this key issue by adopting importance sampling (IS) for local training. We propose importance sampling federated learning (ISFL), an explicit framework with theoretical guarantees. Firstly, we derive the convergence theorem of ISFL to involve the effects of local importance sampling. Then, we formulate the problem of selecting optimal IS weights and obtain the theoretical solutions. We also employ a water-filling method to calculate the IS weights and develop the ISFL algorithms. The experimental results on CIFAR-10 fit the proposed theorems well and verify that ISFL reaps better performance, sampling efficiency, as well as explainability on non-i.i.d. data. To the best of our knowledge, ISFL is the first non-i.i.d. FL solution from the local sampling aspect which exhibits theoretical compatibility with neural network models. Furthermore, as a local sampling approach, ISFL can be easily migrated into other emerging FL frameworks.
翻译:作为将计算和交流相结合的有希望的学习范例,联合会学习(FL)将进行当地培训和定期分享分布客户。由于客户的数据分布不一,FL模型存在梯度多样性、性能差、趋同性差等等。在这项工作中,我们的目标是通过为当地培训采用重要取样(IS)来解决这一关键问题。我们建议取样联合会学习(ISFL)很重要,这是一个有理论保证的明确框架。首先,我们从ISFL的趋同理论理论理论理论中,将地方重要取样的影响纳入地方重要取样中。然后,我们提出选择最佳IS重量和获得理论解决方案的问题。我们还采用填水方法计算ISI重量并开发ISFL算法。CIFAR-10的实验结果符合拟议的标本,并核实ISFL获得更好的业绩、采样效率以及非i.d数据的解释性。据我们所知,ISFL是第一个非i.i.i.然后,我们从当地取样的角度来计算最佳IS IL的重量并获得理论性解决方案。此外,CAR-10的实验结果符合拟议的标准,并核实ISFL在其他采样模型的理论兼容性。</s>