Low-rank and nonsmooth matrix optimization problems capture many fundamental tasks in statistics and machine learning. While significant progress has been made in recent years in developing efficient methods for \textit{smooth} low-rank optimization problems that avoid maintaining high-rank matrices and computing expensive high-rank SVDs, advances for nonsmooth problems have been slow paced. In this paper we consider standard convex relaxations for such problems. Mainly, we prove that under a \textit{strict complementarity} condition and under the relatively mild assumption that the nonsmooth objective can be written as a maximum of smooth functions, approximated variants of two popular \textit{mirror-prox} methods: the Euclidean \textit{extragradient method} and mirror-prox with \textit{matrix exponentiated gradient updates}, when initialized with a "warm-start", converge to an optimal solution with rate $O(1/t)$, while requiring only two \textit{low-rank} SVDs per iteration. Moreover, for the extragradient method we also consider relaxed versions of strict complementarity which yield a trade-off between the rank of the SVDs required and the radius of the ball in which we need to initialize the method. We support our theoretical results with empirical experiments on several nonsmooth low-rank matrix recovery tasks, demonstrating both the plausibility of the strict complementarity assumption, and the efficient convergence of our proposed low-rank mirror-prox variants.
翻译:低调和非高调矩阵优化问题在统计和机器学习方面有许多基本任务。虽然近年来在为避免保持高端矩阵和计算昂贵高端SVD的低级优化问题制定高效方法方面取得了显著进展,但对于非平稳问题的进展速度缓慢。在本文件中,我们考虑对此类问题进行标准的平流放松。主要,我们证明,在\ textit{ 严格互补} 条件下,在相对温和的假设下,非平流目标可以写成一个最顺畅功能的镜像,但两种流行的平流假设的近似变异:Euclidean\ textit{mirror-prox 方法,以及带有\ textitit{matrix Expententented 梯度更新的镜像-prox 。当我们以“温和启动”开始时,我们只需要两个低调的平流的平流的平流数据,与此同时,我们最初的平流的平流的平流的平流方法也需要一些平滑的平流的平流方法。