In this paper, we consider an unconstrained stochastic optimization problem where the objective function exhibits high-order smoothness. Specifically, we propose a new stochastic first-order method (SFOM) with multi-extrapolated momentum, in which multiple extrapolations are performed in each iteration, followed by a momentum update based on these extrapolations. We demonstrate that the proposed SFOM can accelerate optimization by exploiting the high-order smoothness of the objective function $f$. Assuming that the $p$th-order derivative of $f$ is Lipschitz continuous for some $p\ge2$, and under additional mild assumptions, we establish that our method achieves a sample complexity of $\widetilde{\mathcal{O}}(\epsilon^{-(3p+1)/p})$ for finding a point $x$ such that $\mathbb{E}[\|\nabla f(x)\|]\le\epsilon$. To the best of our knowledge, this is the first SFOM to leverage arbitrary-order smoothness of the objective function for acceleration, resulting in a sample complexity that improves upon the best-known results without assuming the mean-squared smoothness condition. Preliminary numerical experiments validate the practical performance of our method and support our theoretical findings.
翻译:本文研究一类无约束随机优化问题,其中目标函数具有高阶光滑性。具体而言,我们提出一种带有多重外推动量的新型随机一阶方法,该方法在每次迭代中执行多次外推操作,随后基于这些外推结果进行动量更新。我们证明所提出的随机一阶方法能够利用目标函数$f$的高阶光滑性来加速优化过程。假设$f$的$p$阶导数($p\ge2$)满足Lipschitz连续性,并在附加的温和假设条件下,我们证明该方法在寻找满足$\mathbb{E}[\|\nabla f(x)\|]\le\epsilon$的点$x$时,能达到$\widetilde{\mathcal{O}}(\epsilon^{-(3p+1)/p})$的样本复杂度。据我们所知,这是首个利用目标函数任意阶光滑性实现加速的随机一阶方法,其样本复杂度在不假设均方光滑条件的情况下改进了现有最佳结果。初步数值实验验证了本方法的实际性能,并支持了理论结论。