We develop an algorithm for parameter-free stochastic convex optimization (SCO) whose rate of convergence is only a double-logarithmic factor larger than the optimal rate for the corresponding known-parameter setting. In contrast, the best previously known rates for parameter-free SCO are based on online parameter-free regret bounds, which contain unavoidable excess logarithmic terms compared to their known-parameter counterparts. Our algorithm is conceptually simple, has high-probability guarantees, and is also partially adaptive to unknown gradient norms, smoothness, and strong convexity. At the heart of our results is a novel parameter-free certificate for SGD step size choice, and a time-uniform concentration result that assumes no a-priori bounds on SGD iterates.
翻译:我们为无参数随机共振优化(SCO)开发了一种算法,其趋同率只是一个比相应的已知参数设置的最佳率更大的双对数系数。 相反,最已知的无参数的SCO无位率是基于在线无参数的遗憾界限,其中含有与已知参数对应的不可避免的超对数术语。我们的算法在概念上很简单,具有高概率保障,并且部分地适应了未知的梯度规范、光滑度和强共性。我们结果的核心是对 SGD 步骤大小选择的新型无参数证书,以及假设SGD 外观没有优先界限的时间统一浓度结果。