When deploying artificial intelligence (AI) in the real world, being able to trust the operation of the AI by characterizing how it performs is an ever-present and important topic. An important and still largely unexplored task in this characterization is determining major factors within the real world that affect the AI's behavior, such as weather conditions or lighting, and either a) being able to give justification for why it may have failed or b) eliminating the influence the factor has. Determining these sensitive factors heavily relies on collected data that is diverse enough to cover numerous combinations of these factors, which becomes more onerous when having many potential sensitive factors or operating in complex environments. This paper investigates methods that discover and separate out individual semantic sensitive factors from a given dataset to conduct this characterization as well as addressing mitigation of these factors' sensitivity. We also broaden remediation of fairness, which normally only addresses socially relevant factors, and widen it to deal with the desensitization of AI with regard to all possible aspects of variation in the domain. The proposed methods which discover these major factors reduce the potentially onerous demands of collecting a sufficiently diverse dataset. In experiments using the road sign (GTSRB) and facial imagery (CelebA) datasets, we show the promise of using this scheme to perform this characterization and remediation and demonstrate that our approach outperforms state of the art approaches.
翻译:当在现实世界部署人工智能(AI)时,在现实世界中部署人工智能(AI)时,能够信任AI的运作,能够通过描述它的表现是一个始终存在和重要的主题,在这一特征中,重要和基本上尚未探索的任务是确定现实世界中影响AI行为的主要因素,例如天气条件或照明,以及(a)能够说明为什么它可能失败的原因,或者(b)消除因素的影响。确定这些敏感因素在很大程度上依赖于收集的数据,这些数据的多样化足以涵盖这些因素的众多组合,而这些因素如果有许多潜在的敏感因素,或在复杂的环境中运作,就会变得更加繁琐。本文调查了发现和分离个体语义敏感因素的方法,这些因素从一个给定的数据集中发现并分离出来,从而减轻这些因素的敏感性。我们还扩大了对公正性的补救范围,通常只处理社会相关因素,并扩大其范围,以处理AI对领域差异的所有可能方面缺乏敏感认识的问题。拟议的方法减少了收集足够多样数据集的潜在繁重需求。在使用道路标志(GSRB)进行实验时,我们用这种图像的状态展示了我们的方法。