Several widely-used first-order saddle-point optimization methods yield an identical continuous-time ordinary differential equation (ODE) that is identical to that of Gradient Descent Ascent (GDA) method when derived naively. However, the convergence properties of these methods are qualitatively different even on simple bilinear games. Thus the ODE perspective, which has proved powerful in analyzing single-objective optimization methods, has not played a similar role in saddle-point optimization. We adopt a framework studied in fluid dynamics -- known as High-Resolution Differential Equations (HRDEs) -- to design differential equation models for several saddle-point optimization methods. Critically, these HRDEs are distinct for various saddle-point optimization methods. Moreover, on bilinear games, the convergence properties of the HRDEs match the qualitative features of the corresponding discrete methods. Additionally, we show that the HRDE of Optimistic Gradient Descent Ascent (OGDA) exhibits \emph{last-iterate convergence} for general monotone variational inequalities. Finally, we provide rates of convergence for the \emph{best-iterate convergence} of the OGDA method, relying solely on the first-order smoothness of the monotone operator.
翻译:几个广泛使用的一阶马鞍优化方法产生一个与精密地从中得来的梯度梯度法(GDA)方法相同的连续时间普通差分方程(ODE),但这些方法的趋同特性在质量上甚至存在于简单的双线游戏中。因此,在分析单一目标优化方法方面已经证明强大的ODE观点在马鞍优化方法方面没有发挥类似的作用。我们采用了一个在流体动态中研究的框架 -- -- 称为高分辨率差异方程式(HRDEs) -- -- 为几种马鞍优化方法设计差异方程模型。关键地说,这些HRDE是不同的。此外,在双线游戏中,HRDEs的趋同特性与相应的离散方法的质量特征相匹配。此外,我们表明,对顶偏的梯度梯度梯度梯度梯度梯度理论(OGDA) 的展览 = emph{Last-iteratate 趋同 用于一般单质差异。最后,我们为单质调调调调调调调调第一/emph{Best-ratateslational-graphetal-grationervementalitydrogy Oraft-godraft-drogardtroductionslations)的趋同率。