Explainable AI (XAI) is widely viewed as a sine qua non for ever-expanding AI research. A better understanding of the needs of XAI users, as well as human-centered evaluations of explainable models are both a necessity and a challenge. In this paper, we explore how HCI and AI researchers conduct user studies in XAI applications based on a systematic literature review. After identifying and thoroughly analyzing 85 core papers with human-based XAI evaluations over the past five years, we categorize them along the measured characteristics of explanatory methods, namely trust, understanding, fairness, usability, and human-AI team performance. Our research shows that XAI is spreading more rapidly in certain application domains, such as recommender systems than in others, but that user evaluations are still rather sparse and incorporate hardly any insights from cognitive or social sciences. Based on a comprehensive discussion of best practices, i.e., common models, design choices, and measures in user studies, we propose practical guidelines on designing and conducting user studies for XAI researchers and practitioners. Lastly, this survey also highlights several open research directions, particularly linking psychological science and human-centered XAI.
翻译:广泛认为可解释的AI(XAI)是不断扩展AI研究的必要条件。更好地了解XAI用户的需要以及对可解释模型进行以人为本的评价既是必要也是挑战。在本文件中,我们探讨了HCI和AI研究人员如何在系统文献审查的基础上对XAI应用进行用户研究。在对过去5年基于人的XAI评估进行明确和彻底分析85份核心文件之后,我们根据解释方法的测量特点,即信任、理解、公平、可使用性和人际关系团队业绩,对85份核心文件进行了分类。我们的研究显示,XAI在某些应用领域,例如推荐系统比其他系统传播得更快,但用户评价仍然很少,几乎没有纳入认知或社会科学的任何见解。在对最佳做法,即共同模式、设计选择和用户研究措施进行全面讨论之后,我们提出了为XAI研究人员和从业人员设计和进行用户研究的实用准则。最后,这项调查还突出了若干开放的研究方向,特别是将心理科学和人际关系联系起来。