Federated Learning (FL) is a distributed machine learning approach where multiple clients work together to solve a machine learning task. One of the key challenges in FL is the issue of partial participation, which occurs when a large number of clients are involved in the training process. The traditional method to address this problem is randomly selecting a subset of clients at each communication round. In our research, we propose a new technique and design a novel regularized client participation scheme. Under this scheme, each client joins the learning process every $R$ communication rounds, which we refer to as a meta epoch. We have found that this participation scheme leads to a reduction in the variance caused by client sampling. Combined with the popular FedAvg algorithm (McMahan et al., 2017), it results in superior rates under standard assumptions. For instance, the optimization term in our main convergence bound decreases linearly with the product of the number of communication rounds and the size of the local dataset of each client, and the statistical term scales with step size quadratically instead of linearly (the case for client sampling with replacement), leading to better convergence rate $\mathcal{O}\left(\frac{1}{T^2}\right)$ compared to $\mathcal{O}\left(\frac{1}{T}\right)$, where $T$ is the total number of communication rounds. Furthermore, our results permit arbitrary client availability as long as each client is available for training once per each meta epoch.
翻译:联邦学习(FL) 是一种分散式的机器学习方法,让多个客户一起合作解决机器学习任务。 FL 的主要挑战之一是部分参与问题,这是当大量客户参与培训过程时发生的部分参与问题。 解决这一问题的传统方法是随机选择每轮通信的客户子集。 在我们的研究中,我们提出了一个新的技术和设计一个全新的正规化客户参与计划。 在这个计划下,每个客户都加入学习过程,每轮我们称之为超时代的每轮通信,我们称之为超时代的。我们发现,这种参与计划导致客户抽样导致差异的减少。结合流行的 FedAvg算法(McMahan等人,2017年),它导致标准假设下的较高比率。例如,我们的主要趋同期与通信回合数量和每个客户本地数据集的大小成线性结合。 统计术语尺度,一是平级规模,而不是线性客户抽样,导致每类客户总趋一致$\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\</s>