Federated Learning has been recently proposed for distributed model training at the edge. The principle of this approach is to aggregate models learned on distributed clients to obtain a new more general "average" model (FedAvg). The resulting model is then redistributed to clients for further training. To date, the most popular federated learning algorithm uses coordinate-wise averaging of the model parameters for aggregation. In this paper, we carry out a complete general mathematical convergence analysis to evaluate aggregation strategies in a federated learning framework. From this, we derive novel aggregation algorithms which are able to modify their model architecture by differentiating client contributions according to the value of their losses. Moreover, we go beyond the assumptions introduced in theory, by evaluating the performance of these strategies and by comparing them with the one of FedAvg in classification tasks in both the IID and the Non-IID framework without additional hypothesis.
翻译:联邦学习联盟最近被提议在边缘进行分布式示范培训。这一方法的原则是汇总在分布式客户方面学到的模型,以获得新的更一般的“平均”模式(FedAvg),由此形成的模型被重新分配给客户接受进一步培训。迄今为止,最受欢迎的联邦学习算法使用协调平均的汇总模型参数。在本文中,我们进行了全面的数学趋同分析,以评价在联合学习框架内的汇总战略。从中,我们得出新的汇总算法,这些算法能够通过根据客户损失的价值区分客户的贡献来修改其模型结构。此外,我们超越了理论中引入的假设,评估这些战略的绩效,并在没有附加假设的情况下,将这些战略与FedAvg在ID和非IID框架中的分类任务进行比较。