Recently, min-max optimization problems have received increasing attention due to their wide range of applications in machine learning (ML). However, most existing min-max solution techniques are either single-machine or distributed algorithms coordinated by a central server. In this paper, we focus on the decentralized min-max optimization for learning with domain constraints, where multiple agents collectively solve a nonconvex-strongly-concave min-max saddle point problem without coordination from any server. Decentralized min-max optimization problems with domain constraints underpins many important ML applications, including multi-agent ML fairness assurance, and policy evaluations in multi-agent reinforcement learning. We propose an algorithm called PRECISION (proximal gradient-tracking and stochastic recursive variance reduction) that enjoys a convergence rate of $O(1/T)$, where $T$ is the maximum number of iterations. To further reduce sample complexity, we propose PRECISION$^+$ with an adaptive batch size technique. We show that the fast $O(1/T)$ convergence of PRECISION and PRECISION$^+$ to an $\epsilon$-stationary point imply $O(\epsilon^{-2})$ communication complexity and $O(m\sqrt{n}\epsilon^{-2})$ sample complexity, where $m$ is the number of agents and $n$ is the size of dataset at each agent. To our knowledge, this is the first work that achieves $O(\epsilon^{-2})$ in both sample and communication complexities in decentralized min-max learning with domain constraints. Our experiments also corroborate the theoretical results.
翻译:最近,由于机器学习(ML)的应用范围很广,微量最大优化问题日益受到重视。然而,大多数现有的微量最大解决方案技术要么是单一机器,要么是由中央服务器协调的分布式算法。在本文件中,我们侧重于分散的微量最大优化,以便以域限制来学习,其中多个代理商在没有任何服务器的协调的情况下,集体解决一个非covex-强力凝固的微量顶点问题。由于域限制,分散的微量最大优化问题成为许多重要的ML应用的基础,包括多剂ML公平保证和多剂强化学习的政策评价。我们提议一个叫做PRECISion(精度梯度跟踪和振动性递增变异性变异性变异性)的算法,即$(T$)是最大变异性。为了进一步降低样本复杂性,我们提议用适应性批量技术来降低微量的微量美元。我们展示了PRECISION和PRECISION-2(美元)的首次趋异性成本, IMRisional_Orentrusal rationalal rodustrational resm rodustrationalal ex resmlational rodustrislationslation rodustrucal $, rolateslationslationslationslationslationslationslational)</s>