Federated Learning (FL) is a novel distributed machine learning which allows thousands of edge devices to train model locally without uploading data concentrically to the server. But since real federated settings are resource-constrained, FL is encountered with systems heterogeneity which causes a lot of stragglers directly and then leads to significantly accuracy reduction indirectly. To solve the problems caused by systems heterogeneity, we introduce a novel self-adaptive federated framework FedSAE which adjusts the training task of devices automatically and selects participants actively to alleviate the performance degradation. In this work, we 1) propose FedSAE which leverages the complete information of devices' historical training tasks to predict the affordable training workloads for each device. In this way, FedSAE can estimate the reliability of each device and self-adaptively adjust the amount of training load per client in each round. 2) combine our framework with Active Learning to self-adaptively select participants. Then the framework accelerates the convergence of the global model. In our framework, the server evaluates devices' value of training based on their training loss. Then the server selects those clients with bigger value for the global model to reduce communication overhead. The experimental result indicates that in a highly heterogeneous system, FedSAE converges faster than FedAvg, the vanilla FL framework. Furthermore, FedSAE outperforms than FedAvg on several federated datasets - FedSAE improves test accuracy by 26.7% and reduces stragglers by 90.3% on average.
翻译:联邦学习组织(FL)是一个新颖的分布式机器学习,它允许数千个边缘设备在不向服务器上传数据的情况下在当地培训模型,而无需以同样的方式向服务器上传数据。但是,由于真正的联邦化设置受到资源限制,因此FL遇到的系统差异性会直接造成许多分解者,然后间接导致大量降低准确性。为了解决系统差异造成的问题,我们引入了一个新型的自我适应联邦化框架FedSAE(FedSAE),这个框架自动调整设备的培训任务,并积极地挑选参与者来缓解性能退化。在这个框架中,我们建议FDEAE(FSAE)利用设备的历史培训任务的全部信息来预测每个设备负担得起的培训工作量。这样,FedSA(FSA)可以估计每个设备的可靠性,并自行调整每轮客户的培训负荷量。 2)我们将我们的框架与积极学习以自我适应的方式选择参与者结合起来。然后,框架通过加速全球模式的融合。在我们的框架中,服务器根据培训的准确性损失来评估设备的价值。 之后,FSA(FDA) 将那些具有更高程度的中央级测试值的客户比FDA(FedA) 更快地选择一个比FedA(V) ASA(ASA) 更快的高级数据框架。