Object detection is a fundamental task in computer vision, which has been greatly progressed through developing large and intricate deep learning models. However, the lack of transparency is a big challenge that may not allow the widespread adoption of these models. Explainable artificial intelligence is a field of research where methods are developed to help users understand the behavior, decision logics, and vulnerabilities of AI-based systems. Black-box explanation refers to explaining decisions of an AI system without having access to its internals. In this paper, we design and implement a black-box explanation method named Black-box Object Detection Explanation by Masking (BODEM) through adopting a new masking approach for AI-based object detection systems. We propose local and distant masking to generate multiple versions of an input image. Local masks are used to disturb pixels within a target object to figure out how the object detector reacts to these changes, while distant masks are used to assess how the detection model's decisions are affected by disturbing pixels outside the object. A saliency map is then created by estimating the importance of pixels through measuring the difference between the detection output before and after masking. Finally, a heatmap is created that visualizes how important pixels within the input image are to the detected objects. The experimentations on various object detection datasets and models showed that BODEM can be effectively used to explain the behavior of object detectors and reveal their vulnerabilities. This makes BODEM suitable for explaining and validating AI based object detection systems in black-box software testing scenarios. Furthermore, we conducted data augmentation experiments that showed local masks produced by BODEM can be used for further training the object detectors and improve their detection accuracy and robustness.
翻译:目标检测是计算机视觉中的一个基本任务,通过开发大型复杂的深度学习模型已经取得了很大的进展。然而,缺乏透明度是一个可能阻碍这些模型普及的重大挑战。可解释人工智能是一个研究领域,开发其中的方法来帮助用户了解基于AI的系统的行为、决策逻辑和弱点。黑盒解释是指在没有访问其内部的情况下解释AI系统决策的方法。在本文中,我们通过采用新的遮挡方法为基于AI的目标检测系统设计和实现了一个黑盒解释方法,称为遮挡式黑盒目标检测解释(BODEM)。我们提出了局部和远程遮挡来生成输入图像的多个版本。局部遮挡用于扰动目标对象内的像素,以确定检测器对这些更改的反应,而远程遮挡用于评估检测模型的决策如何受到扰动对象外的像素的影响。然后通过衡量遮挡前后检测输出之间的差异来估计像素的重要性,从而创建了显要性图。最后,创建一个地图,可视化输入图像中的重要像素对于检测到的对象的重要性。在各种目标检测数据集和模型上的实验表明,BODEM可有效用于解释目标检测器的行为并揭示其弱点。这使BODEM适合解释和验证黑盒软件测试场景中的基于AI的目标检测系统。此外,我们进行的数据增强实验表明,BODEM产生的局部遮挡可以用于进一步训练目标检测器并提高其检测精度和鲁棒性。