The enormous amount of data produced by mobile and IoT devices has motivated the development of federated learning (FL), a framework allowing such devices (or clients) to collaboratively train machine learning models without sharing their local data. FL algorithms (like FedAvg) iteratively aggregate model updates computed by clients on their own datasets. Clients may exhibit different levels of participation, often correlated over time and with other clients. This paper presents the first convergence analysis for a FedAvg-like FL algorithm under heterogeneous and correlated client availability. Our analysis highlights how correlation adversely affects the algorithm's convergence rate and how the aggregation strategy can alleviate this effect at the cost of steering training toward a biased model. Guided by the theoretical analysis, we propose CA-Fed, a new FL algorithm that tries to balance the conflicting goals of maximizing convergence speed and minimizing model bias. To this purpose, CA-Fed dynamically adapts the weight given to each client and may ignore clients with low availability and large correlation. Our experimental results show that CA-Fed achieves higher time-average accuracy and a lower standard deviation than state-of-the-art AdaFed and F3AST, both on synthetic and real datasets.
翻译:移动和 IoT 设备产生的大量数据促进了联合学习(FL)的发展,这一框架允许这些设备(或客户)在不分享其本地数据的情况下合作培训机器学习模型。 FL 算法(如FedAvg)反复累积,由客户在自己的数据集中计算,客户可能表现出不同的参与程度,往往与时间相关并与其他客户相关。本文介绍了FedAvg 类似FL 算法在不同客户可用性和相关客户可用性下首次的趋同分析。我们的分析突出表明,这种关联性如何对算法的趋同率产生不利影响,以及集成战略如何以引导培训转向偏差模式的成本减轻这一影响。我们根据理论分析,提议CA-Fed,一种新的FL算法试图平衡在最大程度趋同速度和尽量减少模型偏差这两个相互矛盾的目标之间取得平衡。为此,CA-Fed 动态调整了给予每个客户的权重度,并可能忽略低可用性和大关联性客户。我们的实验结果表明,CA-Fed 实现了比州- 和合成AD3 和FA- 的合成数据都实现了更高的时间平均准确性和标准偏差。