In this paper, we introduce the tamed stochastic gradient descent method (TSGD) for optimization problems. Inspired by the tamed Euler scheme, which is a commonly used method within the context of stochastic differential equations, TSGD is an explicit scheme that exhibits stability properties similar to those of implicit schemes. As its computational cost is essentially equivalent to that of the well-known stochastic gradient descent method (SGD), it constitutes a very competitive alternative to such methods. We rigorously prove (optimal) sub-linear convergence of the scheme for strongly convex objective functions on an abstract Hilbert space. The analysis only requires very mild step size restrictions, which illustrates the good stability properties. The analysis is based on a priori estimates more frequently encountered in a time integration context than in optimization, and this alternative approach provides a different perspective also on the convergence of SGD. Finally, we demonstrate the usability of the scheme on a problem arising in a context of supervised learning.
翻译:在本文中,我们引入了用于优化问题的有节制的有节制的梯度下降法(TSGD),受有节制的Euler办法启发,这是在有节制的差别方程式中常用的一种方法,TSGD是一个明确的办法,具有与隐含的梯度下降法类似的稳定性特性,其计算成本基本上与众所周知的有节制的梯度下降法(SGD)的计算成本相当,因此它是一种非常有竞争力的替代方法。我们严格地证明(最优)在抽象的Hilbert空间上强烈交融目标函数的子线性组合。这项分析只需要非常温和的步骤大小限制,这说明良好的稳定性特性。这项分析基于在时间整合方面比优化时更经常遇到的事先估计,而这种备选方法也为SGD的趋同提供了不同的观点。最后,我们证明在监督学习过程中出现的问题上,该计划的实用性是可行的。