The heterogeneity of hardware and data is a well-known and studied problem in the community of Federated Learning (FL) as running under heterogeneous settings. Recently, custom-size client models trained with Knowledge Distillation (KD) has emerged as a viable strategy for tackling the heterogeneity challenge. However, previous efforts in this direction are aimed at client model tuning rather than their impact onto the knowledge aggregation of the global model. Despite performance of global models being the primary objective of FL systems, under heterogeneous settings client models have received more attention. Here, we provide more insights into how the chosen approach for training custom client models has an impact on the global model, which is essential for any FL application. We show the global model can fully leverage the strength of KD with heterogeneous data. Driven by empirical observations, we further propose a new approach that combines KD and Learning without Forgetting (LwoF) to produce improved personalised models. We bring heterogeneous FL on pair with the mighty FedAvg of homogeneous FL, in realistic deployment scenarios with dropping clients.
翻译:硬件和数据的异质性是联邦学习联合会(FL)社区中一个众所周知和研究过的一个问题,因为它在多种环境下运行。最近,经过知识蒸馏(KD)培训的定制规模客户模式已成为解决异质性挑战的可行战略。然而,以前朝这一方向作出的努力旨在客户模式调整,而不是影响全球模式的知识汇总。尽管全球模型的绩效是FL系统的首要目标,但在不同环境的客户模式下,客户模式受到更多的关注。在这里,我们更深入地了解为培训定制客户模式所选择的方法如何对全球模式产生影响,而这种模式对于FL的任何应用都至关重要。我们展示了全球模式能够充分利用多样性数据来利用KD的力量。在经验观察的驱动下,我们进一步提出了一种将KD和LwoF)结合而不会忘记(LwoFF)来产生更好的个性化模型的新方法。我们把异性FL配对与精美的FDAvg,在与投放客户的现实部署情景下。