Counterfactual interventions are a powerful tool to explain the decisions of a black-box decision process, and to enable algorithmic recourse. They are a sequence of actions that, if performed by a user, can overturn an unfavourable decision made by an automated decision system. However, most of the current methods provide interventions without considering the user's preferences. For example, a user might prefer doing certain actions with respect to others. In this work, we present the first human-in-the-loop approach to perform algorithmic recourse by eliciting user preferences. We introduce a polynomial procedure to ask choice-set questions which maximize the Expected Utility of Selection (EUS), and use it to iteratively refine our cost estimates in a Bayesian setting. We integrate this preference elicitation strategy into a reinforcement learning agent coupled with Monte Carlo Tree Search for efficient exploration, so as to provide personalized interventions achieving algorithmic recourse. An experimental evaluation on synthetic and real-world datasets shows that a handful of queries allows to achieve a substantial reduction in the cost of interventions with respect to user-independent alternatives.
翻译:反事实干预是解释黑箱决策程序的决定和提供算法追索的有力工具。它们是一系列行动,如果由用户执行,可以推翻自动决定系统作出的不利决定。然而,目前大多数方法提供干预,而没有考虑用户的偏好。例如,用户可能更愿意对他人采取某些行动。在这项工作中,我们提出了第一个通过吸引用户偏好来进行算法追索的“人与人间交往”方法。我们引入了一个多元程序来提出选择设定的问题,以最大限度地提高选择的预期用途(EUS),并利用它来反复地改进我们在巴伊西亚环境中的成本估算。我们把这种优惠吸引战略纳入一个强化学习剂,加上蒙特卡洛树搜索,以有效探索,从而提供个性化干预,从而实现算法追索。对合成和真实世界数据集进行实验性评估后发现,少数查询可以大幅降低与用户独立的替代品有关的干预成本。