Smoothed online learning has emerged as a popular framework to mitigate the substantial loss in statistical and computational complexity that arises when one moves from classical to adversarial learning. Unfortunately, for some spaces, it has been shown that efficient algorithms suffer an exponentially worse regret than that which is minimax optimal, even when the learner has access to an optimization oracle over the space. To mitigate that exponential dependence, this work introduces a new notion of complexity, the generalized bracketing numbers, which marries constraints on the adversary to the size of the space, and shows that an instantiation of Follow-the-Perturbed-Leader can attain low regret with the number of calls to the optimization oracle scaling optimally with respect to average regret. We then instantiate our bounds in several problems of interest, including online prediction and planning of piecewise continuous functions, which has many applications in fields as diverse as econometrics and robotics.
翻译:平滑的在线学习已经成为一个受欢迎的框架,可以减轻从古典学习到对抗性学习时在统计和计算复杂性方面的巨大损失。 不幸的是,对于某些空间来说,事实证明高效的算法比最优化的算法遭遇了比最理想的差得多的遗憾,即使学习者在空间上能够获得最优化的触角。 为了减轻这种指数依赖性,这项工作引入了一种新的复杂性概念,即宽泛的括号数,将对手所受的限制与空间的大小相去甚远,并表明“Perturbed-Leader ” 的即时化可以对最优化或触角的呼声数量产生低的遗憾,以达到平均的遗憾程度。 我们随后在几个令人感兴趣的问题上,包括在线预测和细小的连续功能规划,这些功能在生态计量和机器人等不同领域有许多应用。