Federated Learning (FL) is a collaborative machine learning (ML) framework that combines on-device training and server-based aggregation to train a common ML model among distributed agents. In this work, we propose an asynchronous FL design with periodic aggregation to tackle the straggler issue in FL systems. Considering limited wireless communication resources, we investigate the effect of different scheduling policies and aggregation designs on the convergence performance. Driven by the importance of reducing the bias and variance of the aggregated model updates, we propose a scheduling policy that jointly considers the channel quality and training data representation of user devices. The effectiveness of our channel-aware data-importance-based scheduling policy, compared with state-of-the-art methods proposed for synchronous FL, is validated through simulations. Moreover, we show that an "age-aware" aggregation weighting design can significantly improve the learning performance in an asynchronous FL setting.
翻译:联邦学习联合会(FL)是一个合作机器学习(ML)框架,将设备培训和服务器汇总结合起来,在分布式代理商中培训一个通用 ML 模型。在这项工作中,我们建议采用非同步的FL 设计,定期汇总,以解决FL 系统中的分流问题。考虑到无线通信资源有限,我们调查不同时间安排政策和组合设计对趋同性能的影响。出于减少综合模型更新的偏差和差异的重要性,我们建议一项时间安排政策,共同考虑用户设备的频道质量和培训数据代表。我们基于频道的基于数据进口的时间安排政策与为同步FL提出的最先进的方法相比,其有效性通过模拟得到验证。此外,我们表明“认识到年龄”的组合加权设计可以大大改善在单调式FL 设置中的学习表现。