Ensuring trusted artificial intelligence (AI) in the real world is an critical challenge. A still largely unexplored task is the determination of the major factors within the real world that affect the behavior and robustness of a given AI module (e.g. weather or illumination conditions). Specifically, here we seek to discover the factors that cause AI systems to fail, and to mitigate their influence. The identification of these factors usually heavily relies on the availability of data that is diverse enough to cover numerous combinations of these factors, but the exhaustive collection of this data is onerous and sometimes impossible in complex environments. This paper investigates methods that discover and mitigate the effects of semantic sensitive factors within a given dataset. We also here generalize the definition of fairness, which normally only addresses socially relevant factors, and widen it to deal with -- more broadly -- the desensitization of AI systems with regard to all possible aspects of variation in the domain. The proposed methods which discover these major factors reduce the potentially onerous demands of collecting a sufficiently diverse dataset. In experiments using road sign (GTSRB) and facial imagery (CelebA) datasets, we show the promise of these new methods and show that they outperform state of the art approaches.
翻译:在现实世界中,确保受信任的人工智能(AI)是一个关键的挑战。大部分尚未探索的任务是确定现实世界中影响某一AI模块行为和稳健性的主要因素(例如天气或照明条件)。具体地说,我们在此寻求发现导致AI系统失败的因素,并减轻其影响。这些因素的确定通常主要取决于能否获得足以涵盖这些因素的众多组合的数据,但这种数据的详尽收集在复杂环境中是繁重的,有时甚至是不可能的。本文调查发现和减轻某一数据集中语义敏感因素的影响的方法。我们在此还概括了公平性的定义,通常只涉及社会相关因素,扩大它的范围,以便更广泛地处理AI系统对领域变化所有可能方面缺乏敏感认识的问题。拟议方法发现这些主要因素,减少了收集足够多样化数据集的潜在繁重需求。在使用路标(GTSRB)和面部图像(CelebA)进行实验时,我们展示了这些新方法的前景,并展示了这些新方法的形式。