We study the sequential general online regression, known also as the sequential probability assignments, under logarithmic loss when compared against a broad class of experts. We focus on obtaining tight, often matching, lower and upper bounds for the sequential minimax regret that are defined as the excess loss it incurs over a class of experts. After proving a general upper bound, we consider some specific classes of experts from Lipschitz class to bounded Hessian class and derive matching lower and upper bounds with provably optimal constants. Our bounds work for a wide range of values of the data dimension and the number of rounds. To derive lower bounds, we use tools from information theory (e.g., Shtarkov sum) and for upper bounds, we resort to new "smooth truncated covering" of the class of experts. This allows us to find constructive proofs by applying a simple and novel truncated Bayesian algorithm. Our proofs are substantially simpler than the existing ones and yet provide tighter (and often optimal) bounds.
翻译:我们研究连续的通用在线回归,也称为相继概率分配,与广泛的专家类别相比,在对数损失下进行对数损失。我们注重获得紧凑的、往往匹配的、下下限和上界的顺序迷你鱼遗憾,即它定义为它给某一类专家带来的超重损失。在证明一个总的上界之后,我们考虑到来自Lipschitz 类专家的某些特定类别,将Hessian 类与捆绑起来,并用可辨最佳的常数来匹配下界和上界。我们的底线对着数据维度和子弹数量的广泛值起作用。为了获得更低界,我们使用信息理论(例如Shtarkov summ)和上界的工具,我们使用新的专家类“moth 疏漏覆盖” 。这使我们能够通过应用简单和新颖的疏漏的巴伊西亚算法来找到建设性的证据。我们的证据比现有的标准简单得多,但提供了更紧凑(而且往往是最优化的)约束。