We study the problem of online generalized linear regression in the stochastic setting, where the label is generated from a generalized linear model with possibly unbounded additive noise. We provide a sharp analysis of the classical follow-the-regularized-leader (FTRL) algorithm to cope with the label noise. More specifically, for $\sigma$-sub-Gaussian label noise, our analysis provides a regret upper bound of $O(\sigma^2 d \log T) + o(\log T)$, where $d$ is the dimension of the input vector, $T$ is the total number of rounds. We also prove a $\Omega(\sigma^2d\log(T/d))$ lower bound for stochastic online linear regression, which indicates that our upper bound is nearly optimal. In addition, we extend our analysis to a more refined Bernstein noise condition. As an application, we study generalized linear bandits with heteroscedastic noise and propose an algorithm based on FTRL to achieve the first variance-aware regret bound.
翻译:该论文研究了随机设置下广义线性回归在在线问题中的情况,其中标签来自于广义线性模型,包含可能无界的添加噪声。为此,本研究提供了FTRL算法的详细分析,以解决标签噪声问题。具体而言,针对$\sigma$-sub-Gaussian标签噪声,本文提出的分析可提供遗憾度的上界为$O(\sigma^2d\log T) + o(\log T)$,其中$d$是输入向量的维数,$T$是总的回合数。同时,我们还证明了随机在线线性回归的$\Omega(\sigma^2dlog(T/d))$下限,这表明了我们的上限与最优解几乎相同。此外,我们将我们的分析扩展到更为复杂的伯恩斯坦噪声条件。最后,以广义线性赌博机带有异方差噪声为例,提出了基于FTRL算法的解决方案,该方案可实现最佳的方差感知及遗憾上限。