Counterfactual interventions are a powerful tool to explain the decisions of a black-box decision process and to enable algorithmic recourse. They are a sequence of actions that, if performed by a user, can overturn an unfavourable decision made by an automated decision system. However, most of the current methods provide interventions without considering the user's preferences. In this work, we propose a shift of paradigm by providing a novel formalization which considers the user as an active part of the process rather than a mere target. Following the preference elicitation setting, we introduce the first human-in-the-loop approach to perform algorithmic recourse. We also present a polynomial procedure to ask questions which maximize the Expected Utility of Selection (EUS), a measure of the utility of the choice set that accounts for the uncertainty with respect to both the model and the user response. We use it to iteratively refine our cost estimates in a Bayesian fashion. We integrate this preference elicitation strategy into a reinforcement learning agent coupled with Monte Carlo Tree Search for the efficient exploration, so as to provide personalized interventions achieving algorithmic recourse. An experimental evaluation of synthetic and real-world datasets shows that a handful of queries allows for achieving a substantial reduction in the cost of interventions with respect to user-independent alternatives.
翻译:反事实干预是解释黑箱决策程序的决定和提供算法追索的有力工具。它们是一系列行动,如果由用户执行,可以推翻自动决定系统作出的不利决定。然而,目前大多数方法提供干预,而没有考虑用户的偏好。在这项工作中,我们提出范式转变,提供一种新颖的正规化,将用户视为程序的积极部分,而不是单纯的目标。在偏好调试设置之后,我们引入了第一个进行算法追索的“在行中人”方法。我们还提出了一个多元程序,询问如何最大限度地提高选择的预期效用(EUS)的问题,这是一套选择的效用衡量标准,既说明模型的不确定性,又说明用户的反应。我们用它来反复地改进我们的成本估算,用巴伊西亚的方式。我们将这种偏好吸引战略纳入一个强化学习剂,同时结合蒙特卡洛树搜索高效探索,以便提供实现算法追索的个性化干预措施。对合成和现实世界用户数据设置进行实验性评估,以大幅降低成本。