Adaptive gradient methods such as AdaGrad and its variants update the stepsize in stochastic gradient descent on the fly according to the gradients received along the way; such methods have gained widespread use in large-scale optimization for their ability to converge robustly, without the need to fine-tune the stepsize schedule. Yet, the theoretical guarantees to date for AdaGrad are for online and convex optimization. We bridge this gap by providing theoretical guarantees for the convergence of AdaGrad for smooth, nonconvex functions. We show that the norm version of AdaGrad (AdaGrad-Norm) converges to a stationary point at the $\mathcal{O}(\log(N)/\sqrt{N})$ rate in the stochastic setting, and at the optimal $\mathcal{O}(1/N)$ rate in the batch (non-stochastic) setting -- in this sense, our convergence guarantees are 'sharp'. In particular, the convergence of AdaGrad-Norm is robust to the choice of all hyper-parameters of the algorithm, in contrast to stochastic gradient descent whose convergence depends crucially on tuning the step-size to the (generally unknown) Lipschitz smoothness constant and level of stochastic noise on the gradient. Extensive numerical experiments are provided to corroborate our theory; moreover, the experiments suggest that the robustness of AdaGrad-Norm extends to state-of-the-art models in deep learning, without sacrificing generalization.
翻译:AdaGrad 及其变体等适应性梯度方法更新了根据沿途收到的梯度,在飞翔梯度梯度下降的阶梯化步骤;这些方法在大规模优化中得到了广泛使用,以便在不需要微调进度表的情况下,能够稳健地汇聚。然而,迄今为止AdaGrad 的理论保障是在线和 convex优化。我们通过为AdaGrad 的平稳、非Convex 功能的趋同提供理论保障来弥合这一差距。我们表明,AdaGrad(AdaGrad-Norm) 的常态版(AdaGrad-Norm) 已经逐渐接近于一个固定点,在 $\mathcal{O} (\log(N)/\sqrt{N}) 上,在StagradGrad(非Convelopic) 中,Adagrad-Nord-Norm (Adgrad-Norm) 的常态性模型的趋同性水平, 提供了更精确的更精确的更精确的更精确的比。