Federated learning~(FL) has recently attracted increasing attention from academia and industry, with the ultimate goal of achieving collaborative training under privacy and communication constraints. Existing iterative model averaging based FL algorithms require a large number of communication rounds to obtain a well-performed model due to extremely unbalanced and non-i.i.d data partitioning among different clients. Thus, we propose FedDM to build the global training objective from multiple local surrogate functions, which enables the server to gain a more global view of the loss landscape. In detail, we construct synthetic sets of data on each client to locally match the loss landscape from original data through distribution matching. FedDM reduces communication rounds and improves model quality by transmitting more informative and smaller synthesized data compared with unwieldy model weights. We conduct extensive experiments on three image classification datasets, and results show that our method can outperform other FL counterparts in terms of efficiency and model performance. Moreover, we demonstrate that FedDM can be adapted to preserve differential privacy with Gaussian mechanism and train a better model under the same privacy budget.
翻译:联邦学习-(FL)最近吸引了学术界和产业界越来越多的关注,最终目标是在隐私和通信限制下实现合作培训。现有的基于平均FL算法的迭代模型需要大量的通信周期才能获得一个完善的模型,因为不同客户之间极不平衡和非i.i.d数据分割。因此,我们建议FedDM从多个本地代理功能中建立全球培训目标,使服务器能够对损失场景有更全面的了解。我们详细地构建每个客户的合成数据集,以便与原始数据相比,通过分配匹配,与当地损失情况相匹配。FedDM减少了通信周期,并通过传送信息更加丰富、更小的合成数据来改进模型质量,而与不易变式的模型重量相比。我们对三个图像分类数据集进行了广泛的实验,结果表明,从效率和模型性能上看,我们的方法可以超越FDDDDM能够适应与Gausian机制的不同隐私,并在相同的隐私预算下培训更好的模型。