We prove lower bounds for higher-order methods in smooth non-convex finite-sum optimization. Our contribution is threefold: We first show that a deterministic algorithm cannot profit from the finite-sum structure of the objective, and that simulating a pth-order regularized method on the whole function by constructing exact gradient information is optimal up to constant factors. We further show lower bounds for randomized algorithms and compare them with the best known upper bounds. To address some gaps between the bounds, we propose a new second-order smoothness assumption that can be seen as an analogue of the first-order mean-squared smoothness assumption. We prove that it is sufficient to ensure state-of-the-art convergence guarantees, while allowing for a sharper lower bound.
翻译:我们证明,在平滑的非碳氢化合物有限和优化中,高阶方法的界限较低。我们的贡献有三重:我们首先表明,确定性算法不能从目标的有限和结构中获利,而通过构建精确的梯度信息,模拟整个函数的pth-顺序规范化方法最优于常数因素。我们进一步显示随机化算法的较低界限,并将其与最已知的上限进行比较。为了解决界限之间的某些差距,我们提出了一个新的第二顺序平稳假设,可以被视为第一级平均和均匀假设的类似物。我们证明,这足以确保最先进的趋同保证,同时允许更精确的更低界限。