Interaction and cooperation with humans are overarching aspirations of artificial intelligence (AI) research. Recent studies demonstrate that AI agents trained with deep reinforcement learning are capable of collaborating with humans. These studies primarily evaluate human compatibility through "objective" metrics such as task performance, obscuring potential variation in the levels of trust and subjective preference that different agents garner. To better understand the factors shaping subjective preferences in human-agent cooperation, we train deep reinforcement learning agents in Coins, a two-player social dilemma. We recruit participants for a human-agent cooperation study and measure their impressions of the agents they encounter. Participants' perceptions of warmth and competence predict their stated preferences for different agents, above and beyond objective performance metrics. Drawing inspiration from social science and biology research, we subsequently implement a new "partner choice" framework to elicit revealed preferences: after playing an episode with an agent, participants are asked whether they would like to play the next round with the same agent or to play alone. As with stated preferences, social perception better predicts participants' revealed preferences than does objective performance. Given these results, we recommend human-agent interaction researchers routinely incorporate the measurement of social perception and subjective preferences into their studies.
翻译:与人类的互动和合作是人工智能(AI)研究的首要愿望。最近的研究表明,受过深层强化学习培训的AI代理商能够与人类合作。这些研究主要通过任务表现等“客观”衡量标准评估人类兼容性,这些衡量标准包括任务表现、隐蔽信任程度的潜在差异以及不同代理商获得的主观偏好。为了更好地了解决定人类代理商合作主观偏好的因素,我们在Coins培训深强化学习代理商,这是一个双人社会困境。我们征聘参与者进行人体代理商合作研究,并衡量他们对其所遇到代理商的印象。参与者对温暖和能力的看法预测他们对不同代理商的公开偏好,高于和超出客观绩效衡量标准。从社会科学和生物学研究中汲取灵感,我们随后实施一个新的“伙伴选择”框架,以获得公开的偏好:在与代理商一起玩一小节之后,请与会者问他们是否愿意与同一代理商一起玩下一轮,还是单独玩。与所述偏好,社会观比客观表现更好地预测参与者的偏好。我们建议,人类代理研究人员互动研究人员经常将衡量社会观点和主观偏好于其研究中。