Object detection is a fundamental task in computer vision, which has been greatly progressed through developing large and intricate deep learning models. However, the lack of transparency is a big challenge that may not allow the widespread adoption of these models. Explainable artificial intelligence is a field of research where methods are developed to help users understand the behavior, decision logics, and vulnerabilities of AI-based systems. Black-box explanation refers to explaining decisions of an AI system without having access to its internals. In this paper, we design and implement a black-box explanation method named Black-box Object Detection Explanation by Masking (BODEM) through adopting a new masking approach for AI-based object detection systems. We propose local and distant masking to generate multiple versions of an input image. Local masks are used to disturb pixels within a target object to figure out how the object detector reacts to these changes, while distant masks are used to assess how the detection model's decisions are affected by disturbing pixels outside the object. A saliency map is then created by estimating the importance of pixels through measuring the difference between the detection output before and after masking. Finally, a heatmap is created that visualizes how important pixels within the input image are to the detected objects. The experimentations on various object detection datasets and models showed that BODEM can be effectively used to explain the behavior of object detectors and reveal their vulnerabilities. This makes BODEM suitable for explaining and validating AI based object detection systems in black-box software testing scenarios. Furthermore, we conducted data augmentation experiments that showed local masks produced by BODEM can be used for further training the object detectors and improve their detection accuracy and robustness.
翻译:对象检测是计算机视觉中的一个基本任务,通过开发大型和复杂的深度学习模型已经取得了巨大进展。然而,缺乏透明度是一个重大挑战,可能不允许这些模型的广泛采用。可解释的人工智能是一种研究领域,该领域开发方法帮助用户理解基于人工智能技术的系统的行为,决策逻辑和漏洞。黑盒解释是指在没有访问其内部的情况下解释AI系统的决策。在本文中,我们通过采用一种新的遮盖方法为基于人工智能的对象检测系统设计和实现了一个黑盒解释方法,名为Black-box对象检测说明通过遮盖(BODEM)。我们提出了局部和远程遮罩以生成输入图像的多个版本。局部掩码用于干扰目标对象内部的像素以确定对象检测器对这些更改的反应方式,而远程掩码被用于评估检测模型的决策如何受到干扰目标对象外的像素的影响。接着,通过测量遮罩前后检测输出的差异来估计像素的重要性创建显著性图。最后,创建热力图,可视化输入图像中的重要像素与检测到的对象之间的关系。对各种对象检测数据集和模型的实验表明,BODEM可有效地用于说明对象检测器的行为并揭示其漏洞。这使BODEM适用于在黑盒软件测试场景中解释和验证基于人工智能的对象检测系统。此外,我们进行了数据增强实验,结果表明BODEM生成的局部MASK可用于进一步训练对象检测器,并提高其检测精度和鲁棒性。