Federated Learning (FL) is a distributed learning paradigm that scales on-device learning collaboratively and privately. Standard FL algorithms such as FedAvg are primarily geared towards smooth unconstrained settings. In this paper, we study the Federated Composite Optimization (FCO) problem, in which the loss function contains a non-smooth regularizer. Such problems arise naturally in FL applications that involve sparsity, low-rank, monotonicity, or more general constraints. We first show that straightforward extensions of primal algorithms such as FedAvg are not well-suited for FCO since they suffer from the "curse of primal averaging," resulting in poor convergence. As a solution, we propose a new primal-dual algorithm, Federated Dual Averaging (FedDualAvg), which by employing a novel server dual averaging procedure circumvents the curse of primal averaging. Our theoretical analysis and empirical experiments demonstrate that FedDualAvg outperforms the other baselines.
翻译:联邦学习(FL)是一个分布式的学习模式,它通过合作和私下的方式在设备上推广学习。标准的FedAvg(FedAvg)等FL算法主要针对平滑的不受限制的环境。在本文中,我们研究了联邦综合优化(FCO)问题,其中损失函数包含非抽吸调节器。这类问题自然出现在FL应用程序中,这些应用程序涉及宽度、低级别、单调或更普遍的制约。我们首先显示FedAvg(FedAvg)等原始算法的直截了当的扩展对FCO并不合适,因为它们受到“原始平均诅咒”的影响,导致趋同性差。作为一种解决办法,我们提出了一个新的原始算法,即“FedDualAvg”(FedDualAvg),它使用新型服务器双均率程序绕过原始平均的诅咒。我们的理论分析和经验实验表明FedDualAvg(FedDualAvg)超越了其他基线。