Min-max optimization problems involving nonconvex-nonconcave objectives have found important applications in adversarial training and other multi-agent learning settings. Yet, no known gradient descent-based method is guaranteed to converge to (even local notions of) min-max equilibrium in the nonconvex-nonconcave setting. For all known methods, there exist relatively simple objectives for which they cycle or exhibit other undesirable behavior different from converging to a point, let alone to some game-theoretically meaningful one~\cite{flokas2019poincare,hsieh2021limits}. The only known convergence guarantees hold under the strong assumption that the initialization is very close to a local min-max equilibrium~\cite{wang2019solving}. Moreover, the afore-described challenges are not just theoretical curiosities. All known methods are unstable in practice, even in simple settings. We propose the first method that is guaranteed to converge to a local min-max equilibrium for smooth nonconvex-nonconcave objectives. Our method is second-order and provably escapes limit cycles as long as it is initialized at an easy-to-find initial point. Both the definition of our method and its convergence analysis are motivated by the topological nature of the problem. In particular, our method is not designed to decrease some potential function, such as the distance of its iterate from the set of local min-max equilibria or the projected gradient of the objective, but is designed to satisfy a topological property that guarantees the avoidance of cycles and implies its convergence.
翻译:与非混凝土非非混凝土目标有关的最小最大优化问题在对抗性培训和其他多试剂学习环境中找到了重要的应用。 然而,没有已知的梯度下降法保证在非混凝土(convex-nonconcave)环境下接近(甚至当地概念)最小最大平衡。 对于所有已知方法来说,存在相对简单的目标,它们循环或表现出其他不受欢迎的行为,从相交到某个点,更不用说某些游戏-理论意义上有意义的单方 ⁇ cite{flokas2019poincare,hsieh2021限制}。唯一已知的梯度下降法保证在强烈的趋同假设下维持着一种强烈的渐趋接近于(甚至当地概念)最小平衡的方法。此外,前述挑战不仅仅是理论上的曲线。所有已知方法在实践上都不稳定,即使是在简单的环境中。我们提出的第一种方法可以保证在平稳的不调和不调和不调解目标上达到当地最低平衡,我们的方法是第二级和可分级的,我们最初设计的一种最容易避免的最小的递合金周期,其初始和最容易的递解的递归定的递解方法是其最容易的缩性的一种方法。