In stochastic convex optimization problems, most existing adaptive methods rely on prior knowledge about the diameter bound $D$ when the smoothness or the Lipschitz constant is unknown. This often significantly affects performance as only a rough approximation of $D$ is usually known in practice. Here, we bypass this limitation by combining mirror descent with dual averaging techniques and we show that, under oblivious step-sizes regime, our algorithms converge without any prior knowledge on the parameters of the problem. We introduce three oblivious stochastic algorithms to address different settings. The first algorithm is designed for objectives in relative scale, the second one is an accelerated version tailored for smooth objectives, whereas the last one is for relatively-smooth objectives. All three algorithms work without prior knowledge of the diameter of the feasible set, the Lipschitz constant or smoothness of the objective function. We use these results to revisit the problem of solving large-scale semidefinite programs using randomized first-order methods and stochastic smoothing. We extend our framework to relative scale and demonstrate the efficiency and robustness of our methods on large-scale semidefinite programs.
翻译:在随机凸优化问题中,当光滑性或Lipschitz常数未知时,大多数现有自适应方法依赖于可行集直径界$D$的先验知识。由于实践中通常仅知晓$D$的粗略近似值,这往往显著影响算法性能。本文通过将镜像下降法与对偶平均技术相结合,规避了这一限制,并证明在无先验步长机制下,我们的算法无需任何问题参数先验知识即可收敛。我们提出了三种无先验信息的随机算法以应对不同场景:首种算法针对相对尺度目标函数设计,第二种是针对光滑目标函数的加速版本,第三种则适用于相对光滑目标函数。所有三种算法均无需可行集直径、目标函数Lipschitz常数或光滑性的先验知识。我们运用这些结论重新审视了使用随机一阶方法和随机平滑技术求解大规模半定规划的问题,将框架扩展至相对尺度,并在大规模半定规划问题上验证了方法的效率与鲁棒性。