AI systems have seen significant adoption in various domains. At the same time, further adoption in some domains is hindered by inability to fully trust an AI system that it will not harm a human. Besides the concerns for fairness, privacy, transparency, and explainability are key to developing trusts in AI systems. As stated in describing trustworthy AI "Trust comes through understanding. How AI-led decisions are made and what determining factors were included are crucial to understand." The subarea of explaining AI systems has come to be known as XAI. Multiple aspects of an AI system can be explained; these include biases that the data might have, lack of data points in a particular region of the example space, fairness of gathering the data, feature importances, etc. However, besides these, it is critical to have human-centered explanations that are directly related to decision-making similar to how a domain expert makes decisions based on "domain knowledge," that also include well-established, peer-validated explicit guidelines. To understand and validate an AI system's outcomes (such as classification, recommendations, predictions), that lead to developing trust in the AI system, it is necessary to involve explicit domain knowledge that humans understand and use.
翻译:同时,在一些领域,由于无法完全信任一个不会伤害人类的AI系统,进一步采用AI系统受到阻碍。除了对公平、隐私、透明度和解释性的关切是建立AI系统信任的关键。正如在描述值得信赖的AI“信任通过理解产生,如何作出AI领导的决定和包括哪些决定因素是理解的关键。解释AI系统的子领域已经被称为 XAI。可以解释AI系统的多个方面;其中包括数据可能具有的偏见、在举例空间的特定区域缺乏数据点、数据收集的公平性、特征重要性等等。然而,除了这些以外,至关重要的是要有与决策直接相关的以人为中心的解释,这些解释与一个域专家如何根据“主要知识”做出决策有着相似,其中也包括既定的、经同行审查的明确准则。为了理解和验证AI系统的结果(如分类、建议、预测),从而导致对AI系统形成信任,有必要包含明确的知识,使人类能够理解和使用。