We consider sequential prediction with expert advice when data are generated from distributions varying arbitrarily within an unknown constraint set. We quantify relaxations of the classical i.i.d. assumption in terms of these constraint sets, with i.i.d. sequences at one extreme and adversarial mechanisms at the other. The Hedge algorithm, long known to be minimax optimal in the adversarial regime, was recently shown to be minimax optimal for i.i.d. data. We show that Hedge with deterministic learning rates is suboptimal between these extremes, and present a new algorithm that adaptively achieves the minimax optimal rate of regret with respect to our relaxations of the i.i.d. assumption, and does so without knowledge of the underlying constraint set. We analyze our algorithm using the follow-the-regularized-leader framework, and prove it corresponds to Hedge with an adaptive learning rate that implicitly scales as the square root of the entropy of the current predictive distribution, rather than the entropy of the initial predictive distribution.
翻译:我们认为,如果数据是在一个未知的制约下,从分布上任意生成的,则有专家建议进行顺序预测。我们用这些制约组来量化古典i.d.假设的放松,在一个极端和对立机制中以i.d.d.顺序进行计算。在对抗制中久以迷你最大优化而闻名的格子算法,最近被证明对i.d.数据来说是最优的。我们显示,具有确定学习率的格子在这些极端之间并不理想,我们提出一种新的算法,在适应性地实现与i.i.d.假设的放松有关的最起码的遗憾率,并且在这样做时没有了解基本的制约组。我们用后定型领导框架来分析我们的算法,并证明它与适应性学习率相匹配,后者隐含着当前预测分布的正方根,而不是最初预测分布的方根。