Smart devices, such as smartphones, wearables, robots, and others, can collect vast amounts of data from their environment. This data is suitable for training machine learning models, which can significantly improve their behavior, and therefore, the user experience. Federated learning is a young and popular framework that allows multiple distributed devices to train deep learning models collaboratively while preserving data privacy. Nevertheless, this approach may not be optimal for scenarios where data distribution is non-identical among the participants or changes over time, causing what is known as concept drift. Little research has yet been done in this field, but this kind of situation is quite frequent in real life and poses new challenges to both continual and federated learning. Therefore, in this work, we present a new method, called Concept-Drift-Aware Federated Averaging (CDA-FedAvg). Our proposal is an extension of the most popular federated algorithm, Federated Averaging (FedAvg), enhancing it for continual adaptation under concept drift. We empirically demonstrate the weaknesses of regular FedAvg and prove that CDA-FedAvg outperforms it in this type of scenario.
翻译:智能手机、穿戴器、机器人等智能设备等智能设备可以从环境中收集大量数据。这些数据适合于培训机器学习模式,可以大大改善他们的行为,从而大大改善用户的经验。 联邦学习是一个年轻而受欢迎的框架,允许多种分布式设备在保护数据隐私的同时合作培训深层次学习模式。然而,对于在参与者中数据分配不统一或随着时间变化而造成所谓概念漂移的情况,这一方法可能不是最佳办法。这一领域尚未进行过少量研究,但这种情形在现实生活中非常频繁,对持续学习和联合学习都构成新的挑战。因此,在这项工作中,我们提出了一种新方法,称为“概念-自由-软件-软件-联合变异”(CDA-FedAvg),我们的提议是扩展最受欢迎的 federeration 算法(FedAvg)的延伸,在概念漂移下加强持续适应。我们从经验上展示了常规的 FedAvg的弱点,并证明CDA-FedAvg在这种情景中超越了CDA-Fevg。