Existing federated learning paradigms usually extensively exchange distributed models at a central solver to achieve a more powerful model. However, this would incur severe communication burden between a server and multiple clients especially when data distributions are heterogeneous. As a result, current federated learning methods often require a large number of communication rounds in training. Unlike existing paradigms, we introduce an alternative perspective to significantly decrease the communication cost in federate learning. In this work, we first introduce a meta knowledge representation method that extracts meta knowledge from distributed clients. The extracted meta knowledge encodes essential information that can be used to improve the current model. As the training progresses, the contributions of training samples to a federated model also vary. Thus, we introduce a dynamic weight assignment mechanism that enables samples to contribute adaptively to the current model update. Then, informative meta knowledge from all active clients is sent to the server for model update. Training a model on the combined meta knowledge without exposing original data among different clients can significantly mitigate the heterogeneity issues. Moreover, to further ameliorate data heterogeneity, we also exchange meta knowledge among clients as conditional initialization for local meta knowledge extraction. Extensive experiments demonstrate the effectiveness and efficiency of our proposed method. Remarkably, our method outperforms the state-of-the-art by a large margin (from $74.07\%$ to $92.95\%$) on MNIST with a restricted communication budget (i.e. 10 rounds).
翻译:现有的联邦学习模式通常在中央求解器上广泛交流分布式模式,以实现更强大的模式。然而,这将给服务器和多个客户带来严重的通信负担,特别是在数据分布不一的情况下。因此,目前的联邦学习方法往往需要大量的培训周期。与现有的模式不同,我们引入了一种替代观点,以大幅降低联邦学习的通信成本。在这项工作中,我们首先引入一种从分布式客户中提取元知识的元知识代表方法。提取的元知识将可用于改进当前模式的基本信息编码。随着培训的进展,培训样本对混合型模式的贡献也各不相同。因此,我们引入了一个动态加权分配机制,使样本能够适应当前模式更新。然后,所有活跃客户的丰富元知识被发送到服务器进行模型更新。在不暴露不同客户原始数据的情况下,我们引入一个综合元知识模型模型,可以大大缓解异质性数据问题。此外,我们还在客户之间交换元知识知识知识的初始化和初始化也各不相同。 广泛测试了我们预算效率的大规模模型, 展示了我们预算效率的模型。 广泛测试, 展示了我们预算效率的大规模模型, 展示了一种10-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx