A good model for action-effect prediction, named environment model, is important to achieve sample-efficient decision-making policy learning in many domains like robot control, recommender systems, and patients' treatment selection. We can take unlimited trials with such a model to identify the appropriate actions so that the costs of queries in the real world can be saved. It requires the model to handle unseen data correctly, also called counterfactual data. However, standard data fitting techniques do not automatically achieve such generalization ability and commonly result in unreliable models. In this work, we introduce counterfactual-query risk minimization (CQRM) in model learning for generalizing to a counterfactual dataset queried by a specific target policy. Since the target policies can be various and unknown in policy learning, we propose an adversarial CQRM objective in which the model learns on counterfactual data queried by adversarial policies, and finally derive a tractable solution GALILEO. We also discover that adversarial CQRM is closely related to the adversarial model learning, explaining the effectiveness of the latter. We apply GALILEO in synthetic tasks and a real-world application. The results show that GALILEO makes accurate predictions on counterfactual data and thus significantly improves policies in real-world testing.
翻译:行动效果预测的良好模式,称为环境模型,对于在机器人控制、推荐人系统和病人治疗选择等许多领域实现抽样有效的决策政策学习十分重要。我们可以用这种模式进行无限制的试验,以确定适当的行动,从而节省真实世界的查询费用。它要求模型正确处理无形数据,也称为反事实数据。但是,标准数据适应技术并不自动实现这种概括化能力,通常导致不可靠的模型。在这项工作中,我们引入反事实质风险最小化(CQRM),作为将反事实数据集推广到特定目标政策查询的模型学习。由于目标政策学习可能多种多样,而且未知,因此我们提出对抗性CQRM目标,即模型学习对抗性数据,最后得出一个可移植的解决方案GALILEO。我们还发现,对抗性CQRM与对抗性模型学习密切相关,解释后者的有效性。我们在合成任务和实际应用中应用GALILEO软件。结果显示,GALILEO在现实世界应用中进行精确的预测。