Bilevel Optimization has witnessed notable progress recently with new emerging efficient algorithms and has been applied to many machine learning tasks such as data cleaning, few-shot learning, and neural architecture search. However, little attention has been paid to solve the bilevel problems under distributed setting. Federated learning (FL) is an emerging paradigm which solves machine learning tasks over distributed-located data. FL problems are challenging to solve due to the heterogeneity and communication bottleneck. However, it is unclear how these challenges will affect the convergence of Bilevel Optimization algorithms. In this paper, we study Federated Bilevel Optimization problems. Specifically, we first propose the FedBiO, a deterministic gradient-based algorithm and we show it requires $O(\epsilon^{-2})$ number of iterations to reach an $\epsilon$-stationary point. Then we propose FedBiOAcc to accelerate FedBiO with the momentum-based variance-reduction technique under the stochastic scenario. We show FedBiOAcc has complexity of $O(\epsilon^{-1.5})$. Finally, we validate our proposed algorithms via the important Fair Federated Learning task. More specifically, we define a bilevel-based group fair FL objective. Our algorithms show superior performances compared to other baselines in numerical experiments.
翻译:双级优化最近随着新的高效算法的出现而取得了显著进展,并应用于许多机器学习任务,例如数据清理、微小的学习和神经结构搜索。然而,对于在分布式环境中解决双层问题没有多少注意。联邦学习(FL)是一个新出现的模式,解决对分布式数据的机器学习任务。FL问题由于差异性和沟通瓶颈而难以解决。然而,这些挑战将如何影响双级优化算法的趋同。在本文件中,我们研究了双级优化算法问题。具体地说,我们首先提议FDBIOO,一种基于确定性梯度的梯度算法,我们显示它需要美元(eepsilon ⁇ -2})的迭代数才能达到分配式数据分布式数据的机器学习任务。然后,我们提议FBIOACcc将基于动力的降低差异技术加速FDBIO。我们展示了FBOACC的复杂程度, 具体地说,我们提议的FIAxalalalal 级标准,我们提出了一个重要比O(eplon_BAR_BAR_BAR_BAxalalal) 任务。最后,我们提出了一个比较重要的比额级的高级标准。