We consider the off-policy evaluation (OPE) problem in contextual bandits, where the goal is to estimate the value of a target policy using the data collected by a logging policy. Most popular approaches to the OPE are variants of the doubly robust (DR) estimator obtained by combining a direct method (DM) estimator and a correction term involving the inverse propensity score (IPS). Existing algorithms primarily focus on strategies to reduce the variance of the DR estimator arising from large IPS. We propose a new approach called the Doubly Robust with Information borrowing and Context-based switching (DR-IC) estimator that focuses on reducing both bias and variance. The DR-IC estimator replaces the standard DM estimator with a parametric reward model that borrows information from the 'closer' contexts through a correlation structure that depends on the IPS. The DR-IC estimator also adaptively interpolates between this modified DM estimator and a modified DR estimator based on a context-specific switching rule. We give provable guarantees on the performance of the DR-IC estimator. We also demonstrate the superior performance of the DR-IC estimator compared to the state-of-the-art OPE algorithms on a number of benchmark problems.
翻译:我们考虑了背景土匪的离政策评价问题,目标是利用伐木政策收集的数据估计目标政策的价值。对目标政策最受欢迎的方法,是通过结合直接方法(DM)估计仪和涉及反偏向性分数(IPS)的校正术语而获得的双强估计仪的变体。现有的算法主要侧重于减少大型IPS产生的DR估计仪差异的战略。我们提出了一种新办法,称为“Doubly Robust 与信息借款和基于背景的转换(DR-IC)估计仪,重点是减少偏差和差异。DMS-IC估计仪将标准估计数替换为“DM(DM)估计数”的参数性能模型,通过取决于IPS的关联结构从“Clober”背景中借用信息。DR-IC估计仪的测算仪和基于特定背景的转换规则(DR)调算码(DR-IC)的调控器。我们用SDRAA标准向国家测试基准数字的性能保证。