Asynchronous learning protocols have regained attention lately, especially in the Federated Learning (FL) setup, where slower clients can severely impede the learning process. Herein, we propose \texttt{AsyncDrop}, a novel asynchronous FL framework that utilizes dropout regularization to handle device heterogeneity in distributed settings. Overall, \texttt{AsyncDrop} achieves better performance compared to state of the art asynchronous methodologies, while resulting in less communication and training time overheads. The key idea revolves around creating ``submodels'' out of the global model, and distributing their training to workers, based on device heterogeneity. We rigorously justify that such an approach can be theoretically characterized. We implement our approach and compare it against other asynchronous baselines, both by design and by adapting existing synchronous FL algorithms to asynchronous scenarios. Empirically, \texttt{AsyncDrop} reduces the communication cost and training time, while matching or improving the final test accuracy in diverse non-i.i.d. FL scenarios.
翻译:最近,特别是在联邦学习(FL)的设置中,较慢的客户会严重妨碍学习过程。在这里,我们提出\ texttt{AsyncDrop},这是一个新的非同步FL框架,利用辍学规范处理分布环境中的装置异质性。总体来说,\ textt{AsyncDrop}取得了比艺术非同步状态方法更好的性能,同时减少了沟通和培训时间。关键思想围绕在全球模型中创建“sub model's”,并根据装置异质性向工人分配培训。我们严格证明这样一种方法在理论上是可以定性的。我们采用我们的方法并将其与其他非同步基线进行比较,既通过设计,又通过调整现有同步的FL算法来适应不同步情景。简化,\texttt{Asyncrop}降低了通信成本和培训时间,同时匹配或改进了多种非F-L.i.