We show how to construct variance-aware confidence sets for linear bandits and linear mixture Markov Decision Process (MDP). Our method yields the following new regret bounds: * For linear bandits, we obtain an $\widetilde{O}(\mathrm{poly}(d)\sqrt{1 + \sum_{i=1}^{K}\sigma_i^2})$ regret bound, where $d$ is the feature dimension, $K$ is the number of rounds, and $\sigma_i^2$ is the (unknown) variance of the reward at the $i$-th round. This is the first regret bound that only scales with the variance and the dimension, with no explicit polynomial dependency on $K$. * For linear mixture MDP, we obtain an $\widetilde{O}(\mathrm{poly}(d, \log H)\sqrt{K})$ regret bound, where $d$ is the number of base models, $K$ is the number of episodes, and $H$ is the planning horizon. This is the first regret bound that only scales logarithmically with $H$ in the reinforcement learning with linear function approximation setting, thus exponentially improving existing results. Our methods utilize three novel ideas that may be of independent interest: 1) applications of the peeling techniques to the norm of input and the magnitude of variance, 2) a recursion-based approach to estimate the variance, and 3) a convex potential lemma that somewhat generalizes the seminal elliptical potential lemma.
翻译:我们展示了如何为线性土匪和线性混合物 Markov 决断进程构建有差异的信任度。 我们的方法产生了以下新的遗憾界限 : * 对于线性土匪,我们获得的是全色土匪( mathrm{poly})( d)\qrt{1 +\ sum ⁇ i=1 ⁇ K ⁇ sgma_i ⁇ 2}) 美元约束, 美元是特征维度, 美元是周期性, 美元是周期性美元报酬的( 未知) 。 这是第一个遗憾界限, 只有差异和层面的尺度, 没有明确的多色依赖 $. * 对于线性混合物, 我们得到的是 $\\ 百度{Kgma_ =1\ kgma_ =1\ kgma_ i2} ( poly} ( d, log H) = sqraltial } legleglegle, 美元是基础型模型的数值, $K$x 和直线性模型的模型的值 值 值 。