Much of the literature on optimal design of bandit algorithms is based on minimization of expected regret. It is well known that designs that are optimal over certain exponential families can achieve expected regret that grows logarithmically in the number of arm plays, at a rate governed by the Lai-Robbins lower bound. In this paper, we show that when one uses such optimized designs, the regret distribution of the associated algorithms necessarily has a very heavy tail, specifically, that of a truncated Cauchy distribution. Furthermore, for $p>1$, the $p$'th moment of the regret distribution grows much faster than poly-logarithmically, in particular as a power of the total number of arm plays. We show that optimized UCB bandit designs are also fragile in an additional sense, namely when the problem is even slightly mis-specified, the regret can grow much faster than the conventional theory suggests. Our arguments are based on standard change-of-measure ideas, and indicate that the most likely way that regret becomes larger than expected is when the optimal arm returns below-average rewards in the first few arm plays, thereby causing the algorithm to believe that the arm is sub-optimal. To alleviate the fragility issues exposed, we show that UCB algorithms can be modified so as to ensure a desired degree of robustness to mis-specification. In doing so, we also provide a sharp trade-off between the amount of UCB exploration and the tail exponent of the resulting regret distribution.
翻译:有关最优化设计土匪算法的大量文献都是基于尽量减少预期的遗憾。 众所周知, 最优于某些指数式家庭的最佳设计能够实现预期的遗憾,以拉伊-罗宾斯较低的约束范围,在比例上,使手臂游戏的数量成对地增长。 在本文中,我们表明,当人们使用这种优化设计时,相关算法的遗憾分布必然有一个非常沉重的尾巴,具体地说,就是松散的灰尘分布。此外,对于美元 > 1美元来说,最优于某些指数式家庭的最佳设计可以实现预期的遗憾分配时刻,比多元圆形家庭快得多,特别是作为手臂游戏总数的力量。我们表明,优化的UCB波型组合设计在另一个意义上也很脆弱,也就是说,当问题被略微错误描述时,其遗憾分布会比常规理论要快得多。 我们的论点是以标准的改变计量概念为基础,并且表明最可能后悔的方式是,当最优的手臂在最初几个手臂游戏中返回低于平均的奖赏时,特别是作为总数量的力量。 因此,优化的UCB型组合的设计设计会让我们相信, 能够降低一个稳定度。