We advance a novel computational model of multi-agent, cooperative joint actions that is grounded in the cognitive framework of active inference. The model assumes that to solve a joint task, such as pressing together a red or blue button, two (or more) agents engage in a process of interactive inference. Each agent maintains probabilistic beliefs about the goal of the joint task (e.g., should we press the red or blue button?) and updates them by observing the other agent's movements, while in turn selecting movements that make his own intentions legible and easy to infer by the other agent (i.e., sensorimotor communication). Over time, the interactive inference aligns both the beliefs and the behavioral strategies of the agents, hence ensuring the success of the joint action. We exemplify the functioning of the model in two simulations. The first simulation illustrates a ''leaderless'' joint action. It shows that when two agents lack a strong preference about their joint task goal, they jointly infer it by observing each other's movements. In turn, this helps the interactive alignment of their beliefs and behavioral strategies. The second simulation illustrates a "leader-follower" joint action. It shows that when one agent ("leader") knows the true joint goal, it uses sensorimotor communication to help the other agent ("follower") infer it, even if doing this requires selecting a more costly individual plan. These simulations illustrate that interactive inference supports successful multi-agent joint actions and reproduces key cognitive and behavioral dynamics of "leaderless" and "leader-follower" joint actions observed in human-human experiments. In sum, interactive inference provides a cognitively inspired, formal framework to realize cooperative joint actions and consensus in multi-agent systems.
翻译:我们推进一个基于积极推断的认知框架的多试剂、合作联合行动的新计算模型。 模型假设要完成一个联合任务, 比如按一个红色或蓝色按钮, 两个(或更多)代理人参与一个互动推断过程。 每个代理人都保持对联合任务目标的概率性信念( 比如, 我们是否按下红色或蓝色按钮? ), 并通过观察其他代理人的动作来更新它们, 而反过来选择一些运动, 让另一个代理人( 即感官模型通信) 能够理解自己的意图, 容易推断出另一个代理人( 感官模型通信) 。 随着时间的推移, 互动推断使代理人的信念和行为战略一致, 从而确保联合行动的成功。 我们通过两次模拟来展示模型的功能。 第一次模拟显示“ 缺乏领导力的” 联合行动。 当两个代理人对选择联合任务目标缺乏强烈的偏好, 他们通过观察对方的动作来共同推导出。 反过来, 反过来, 甚至这帮助互动调整了代理人的信念和行为战略, 从而显示, “ 联合动作的动作的比, 当它需要一个真正的代理人的动作和感官的动作时, 。