Data-heterogeneous federated learning (FL) systems suffer from two significant sources of convergence error: 1) client drift error caused by performing multiple local optimization steps at clients, and 2) partial client participation error caused by the fact that only a small subset of the edge clients participate in every training round. We find that among these, only the former has received significant attention in the literature. To remedy this, we propose FedVARP, a novel variance reduction algorithm applied at the server that eliminates error due to partial client participation. To do so, the server simply maintains in memory the most recent update for each client and uses these as surrogate updates for the non-participating clients in every round. Further, to alleviate the memory requirement at the server, we propose a novel clustering-based variance reduction algorithm ClusterFedVARP. Unlike previously proposed methods, both FedVARP and ClusterFedVARP do not require additional computation at clients or communication of additional optimization parameters. Through extensive experiments, we show that FedVARP outperforms state-of-the-art methods, and ClusterFedVARP achieves performance comparable to FedVARP with much less memory requirements.
翻译:为了纠正这一点,我们提议FedVARP在服务器上采用新的差异减少算法,消除部分客户参与导致的错误。为了做到这一点,我们提议FedVARP在服务器上采用新的差异减少算法,消除部分客户参与造成的错误。为了做到这一点,服务器只需保存每个客户的最新更新资料,并在每一轮中作为未参加客户的代理更新资料使用。此外,为了减轻服务器的存储要求,我们提议采用新的基于集群的减少差异算法GroupFedVARP。与先前提议的方法不同,FedVARP和GroupFedVARP都不需要对客户进行额外计算或交流额外的优化参数。通过广泛的实验,我们显示FedVARP超越了最先进的方法,而GrodFedVARP实现了类似于FedVARP的存储要求。