Large scale convex-concave minimax problems arise in numerous applications, including game theory, robust training, and training of generative adversarial networks. Despite their wide applicability, solving such problems efficiently and effectively is challenging in the presence of large amounts of data using existing stochastic minimax methods. We study a class of stochastic minimax methods and develop a communication-efficient distributed stochastic extragradient algorithm, LocalAdaSEG, with an adaptive learning rate suitable for solving convex-concave minimax problem in the Parameter-Server model. LocalAdaSEG has three main features: (i) periodic communication strategy reduces the communication cost between workers and the server; (ii) an adaptive learning rate that is computed locally and allows for tuning-free implementation; and (iii) theoretically, a nearly linear speed-up with respect to the dominant variance term, arising from estimation of the stochastic gradient, is proven in both the smooth and nonsmooth convex-concave settings. LocalAdaSEG is used to solve a stochastic bilinear game, and train generative adversarial network. We compare LocalAdaSEG against several existing optimizers for minimax problems and demonstrate its efficacy through several experiments in both the homogeneous and heterogeneous settings.
翻译:在众多应用中,包括游戏理论、强力培训以及基因对抗网络培训等应用中,出现了大型相形相形细形细形的微缩成交问题。尽管这些问题具有广泛适用性,但是,在使用现有随机小成形方法提供大量数据的情况下,有效、高效地解决这些问题具有挑战性。我们研究了一组随机小型成交方法,并开发了一种具有通信效率的分布式相色相色谱超梯度算法(LocalAdaSEG),该算法具有适应性学习率,适合于解决Parameter-Server模型中的相形色色色谱小型成交织问题。本地AdaSEG有三个主要特征:(一)定期通信战略降低了工人与服务器之间的通信成本;(二)适应性学习率,在本地进行计算,便于调整无限制的实施;以及(三)理论上,由于对色相梯度梯度梯度的估算,对主要差异术语的近线性超速率,在光度和非摩相色相色色相对等相对等组合结构中得到了证明。当地AdaSEGEng用来解决现有最佳双色色色色色色色变游戏和基因实验网络。