Given an increasing prevalence of intelligent systems capable of autonomous actions or augmenting human activities, it is important to consider scenarios in which the human, autonomous system, or both can exhibit failures as a result of one of several contributing factors (e.g. perception). Failures for either humans or autonomous agents can lead to simply a reduced performance level, or a failure can lead to something as severe as injury or death. For our topic, we consider the hybrid human-AI teaming case where a managing agent is tasked with identifying when to perform a delegation assignment and whether the human or autonomous system should gain control. In this context, the manager will estimate its best action based on the likelihood of either (human, autonomous) agent failure as a result of their sensing capabilities and possible deficiencies. We model how the environmental context can contribute to, or exacerbate, the sensing deficiencies. These contexts provide cases where the manager must learn to attribute capabilities to suitability for decision-making. As such, we demonstrate how a Reinforcement Learning (RL) manager can correct the context-delegation association and assist the hybrid team of agents in outperforming the behavior of any agent working in isolation.
翻译:随着越来越多的智能系统能够自主行动或增强人类活动的能力,考虑到人类、自主系统或两者都可能由于多种因素(如感知)而出现失败的情况显得越来越重要。无论是人类还是自主代理出现故障都可能导致性能降低,甚至是严重的伤害或死亡。对于我们的研究主题,我们考虑到了混合人工智能团队工作的情况,其中一个管理代理被要求识别何时执行委派任务,以及是人类还是自主系统应该获得控制权。在这种情况下,管理代理将基于它们感知能力和潜在缺陷的可能性,估计出最佳的行动。我们建立了环境情境的模型,可以增加或加剧感知缺陷,这些情境让管理者必须学习将决策能力归因于宜于决策的能力。因此,我们展示了如何通过强化学习(RL)管理器来修正上下文-委派关联,以帮助混合代理团队表现超过任何单独工作的代理。