Recent work has shown that, in classification tasks, it is possible to design decision support systems that do not require human experts to understand when to cede agency to a classifier or when to exercise their own agency to achieve complementarity$\unicode{x2014}$experts using these systems make more accurate predictions than those made by the experts or the classifier alone. The key principle underpinning these systems reduces to adaptively controlling the level of human agency, by design. Can we use the same principle to achieve complementarity in sequential decision making tasks? In this paper, we answer this question affirmatively. We develop a decision support system that uses a pre-trained AI agent to narrow down the set of actions a human can take to a subset, and then asks the human to take an action from this action set. Along the way, we also introduce a bandit algorithm that leverages the smoothness properties of the action sets provided by our system to efficiently optimize the level of human agency. To evaluate our decision support system, we conduct a large-scale human subject study ($n = 1{,}600$) where participants play a wildfire mitigation game. We find that participants who play the game supported by our system outperform those who play on their own by $\sim$$30$% and the AI agent used by our system by $>$$2$%, even though the AI agent largely outperforms participants playing without support. We have made available the data gathered in our human subject study as well as an open source implementation of our system at https://github.com/Networks-Learning/narrowing-action-choices .
翻译:近期研究表明,在分类任务中,可以设计出无需人类专家理解何时将决策权让渡给分类器、何时行使自主权即可实现互补性的决策支持系统$\unicode{x2014}$使用这些系统的专家能做出比单独由专家或分类器更准确的预测。支撑这些系统的核心原理在于通过设计自适应地控制人类决策权水平。我们能否运用相同原理在序列决策任务中实现互补性?本文对此问题给出肯定回答。我们开发了一种决策支持系统,该系统使用预训练的AI智能体将人类可采取的行动集合缩小至子集,随后要求人类从该行动集中选择行动。在此过程中,我们还提出了一种多臂赌博机算法,该算法利用系统提供的行动集平滑特性来高效优化人类决策权水平。为评估该决策支持系统,我们开展了大规模人类受试者研究($n = 1{,}600$),参与者需进行野火防控博弈实验。研究发现:在系统支持下参与游戏的受试者表现,较独立游戏者提升约$30$%,较系统使用的AI智能体提升超过$2$%,尽管该AI智能体在无支持条件下显著优于人类参与者。我们已在https://github.com/Networks-Learning/narrowing-action-choices 公开了人类受试者研究数据及系统的开源实现。