Mirror descent, introduced by Nemirovski and Yudin in the 1970s, is a primal-dual convex optimization method that can be tailored to the geometry of the optimization problem at hand through the choice of a strongly convex potential function. It arises as a basic primitive in a variety of applications, including large-scale optimization, machine learning, and control. This paper proposes a variational formulation of mirror descent and of its stochastic variant, mirror Langevin dynamics. The main idea, inspired by the classic work of Brezis and Ekeland on variational principles for gradient flows, is to show that mirror descent emerges as a closed-loop solution for a certain optimal control problem, and the Bellman value function is given by the Bregman divergence between the initial condition and the global minimizer of the objective function.
翻译:由Nemirovski和Yudin在1970年代推出的镜状下坠是一种原始的双向二次曲线优化方法,可以通过选择强烈的曲线潜在功能,适应手头优化问题的几何性。它作为一种基本原始形式,在各种应用中产生,包括大规模优化、机器学习和控制。本文提出了镜状下坠及其随机变异的变体,即镜子Langevin动态。主要思想来自布雷兹和埃克兰关于梯度流动变异原则的经典著作,其灵感是显示镜状下坠是某种最佳控制问题的一种闭路办法,而贝尔曼的价值功能是由布雷格曼在最初条件与目标功能的全球最小化者之间存在的差异提供的。</s>