We propose the homotopic policy mirror descent (HPMD) method for solving discounted, infinite horizon MDPs with finite state and action space, and study its policy convergence. We report several properties that seem to be new in the literature of policy gradient methods: (1) HPMD exhibits global linear convergence of the value optimality gap, and local superlinear convergence of the policy to the set of optimal policies with order $\gamma^{-2}$. The superlinear convergence of the policy takes effect after no more than $\mathcal{O}(\log(1/\Delta^*))$ number of iterations, where $\Delta^*$ is defined via a gap quantity associated with the optimal state-action value function; (2) HPMD also exhibits last-iterate convergence of the policy, with the limiting policy corresponding exactly to the optimal policy with the maximal entropy for every state. No regularization is added to the optimization objective and hence the second observation arises solely as an algorithmic property of the homotopic policy gradient method; (3) The last-iterate convergence of HPMD holds for a much broader class of decomposable distance-generating functions, including the $p$-th power of $\ell_p$-norm and the negative Tsallis entropy. As a byproduct of the analysis, we also discover the finite-time exact convergence of HPMD with these divergences, and show that HPMD continues converging to the limiting policy even if the current policy is already optimal; (4) For the stochastic HPMD method, we further demonstrate that a better than $\mathcal{O}(|\mathcal{S}| |\mathcal{A}| / \epsilon^2)$ sample complexity for small optimality gap $\epsilon$ holds with high probability, when assuming a generative model for policy evaluation.
翻译:我们建议采用同质政策镜底(HPMD)方法来解决有限制的状态和行动空间的折价、无限地平地 MDP(HPMD)方法,并研究其政策趋同。我们报告了一些在政策梯度方法文献中似乎是新的属性:(1) HPMD展示了价值最佳差值的全球线性趋同,以及该政策与一套最佳政策在当地的超线性趋同,以$\gammacha=2美元为主。没有将政策与优化目标相加,因此,该政策的超级线性趋同作用在不超过美元(mathcal{O}(\log(1/Delta ⁇ ))美元乘以迭代数的迭代数后才会生效,其中,$Delta $已经通过与最佳州-行动值值相挂钩的差额来界定。 (2) HPMDMD(Delta)还展示了政策的最后一级趋同性趋同性, 也就是以美元平价变现的变现政策函数,包括变现的变现的变现的变现的变现的变变变变变变变的变的变的变的变的变变的变的变的变的变的变现。