We investigate the problem of explainability for visual object detectors. Specifically, we demonstrate on the example of the YOLO object detector how to integrate Grad-CAM into the model architecture and analyze the results. We show how to compute attribution-based explanations for individual detections and find that the normalization of the results has a great impact on their interpretation.
翻译:我们调查视觉物体探测器的可解释性问题。 具体地说,我们用YOLO物体探测器的例子来示范如何将Grad-CAM纳入模型结构并分析结果。我们展示了如何计算单个探测的归属解释,并发现结果正常化对其解释有重大影响。