Large scale convex-concave minimax problems arise in numerous applications, including game theory, robust training, and training of generative adversarial networks. Despite their wide applicability, solving such problems efficiently and effectively is challenging in the presence of large amounts of data using existing stochastic minimax methods. We study a class of stochastic minimax methods and develop a communication-efficient distributed stochastic extragradient algorithm, LocalAdaSEG, with an adaptive learning rate suitable for solving convex-concave minimax problems in the Parameter-Server model. LocalAdaSEG has three main features: (i) a periodic communication strategy that reduces the communication cost between workers and the server; (ii) an adaptive learning rate that is computed locally and allows for tuning-free implementation; and (iii) theoretically, a nearly linear speed-up with respect to the dominant variance term, arising from the estimation of the stochastic gradient, is proven in both the smooth and nonsmooth convex-concave settings. LocalAdaSEG is used to solve a stochastic bilinear game, and train a generative adversarial network. We compare LocalAdaSEG against several existing optimizers for minimax problems and demonstrate its efficacy through several experiments in both homogeneous and heterogeneous settings.
翻译:在众多应用中,包括游戏理论、强力培训以及基因对抗网络培训等应用中,都出现了大型相交小型算法问题。尽管这些问题具有广泛适用性,但是,在使用现有随机小型算法提供大量数据的情况下,有效、高效地解决这些问题具有挑战性。我们研究了一组随机小型算法,并开发了一个具有通信效率的分布式相交超梯度算法(LocalAdaSEG),该算法的适应性学习率适合于解决参数-服务器模型中的相交小型算法问题。本地AdaSEG有三大特征:(一) 定期通信战略,降低工人与服务器之间的通信成本;(二) 适应性学习率,由本地计算,并允许不作调整地执行;以及(三) 从理论上讲,由于对随机梯度的估算,在主要差异术语方面几乎直线性地提高速度,在光滑和非摩式组合连接式组合环境中都证明了这一点。本地AdaSEGEG用来解决当前数个对抗性双向型智能游戏和Syalimal-Agyalal Intra 。