Nowadays, deep vision models are being widely deployed in safety-critical applications, e.g., autonomous driving, and explainability of such models is becoming a pressing concern. Among explanation methods, counterfactual explanations aim to find minimal and interpretable changes to the input image that would also change the output of the model to be explained. Such explanations point end-users at the main factors that impact the decision of the model. However, previous methods struggle to explain decision models trained on images with many objects, e.g., urban scenes, which are more difficult to work with but also arguably more critical to explain. In this work, we propose to tackle this issue with an object-centric framework for counterfactual explanation generation. Our method, inspired by recent generative modeling works, encodes the query image into a latent space that is structured in a way to ease object-level manipulations. Doing so, it provides the end-user with control over which search directions (e.g., spatial displacement of objects, style modification, etc.) are to be explored during the counterfactual generation. We conduct a set of experiments on counterfactual explanation benchmarks for driving scenes, and we show that our method can be adapted beyond classification, e.g., to explain semantic segmentation models. To complete our analysis, we design and run a user study that measures the usefulness of counterfactual explanations in understanding a decision model. Code is available at https://github.com/valeoai/OCTET.
翻译:目前,深视模型正在安全关键应用中广泛应用,例如自主驱动,而且这些模型的解释性正在成为一个紧迫的关注问题。在解释方法中,反事实解释旨在找到对输入图像的最小和可解释的变化,这种变化也会改变模型的输出。这种解释点指出最终用户是影响模型决定的主要因素。然而,以往的方法在解释关于许多物体的图像(例如,城市景点)所训练的决策模型方面困难重重,这些图像更难与城市景点合作,但可以说更难解释。在这项工作中,我们提议用一个以物体为中心的框架来处理这一问题,以产生反事实解释。我们的方法,在近期的基因化模型的启发下,将查询图像编码成一个潜在的空间,以缓解对物体操作的操纵。这样做,为终端用户提供了控制,在反事实生成过程中将探索方向(例如,物体的空间变异模型,样式的修改等等)。我们用一个反事实解释基准来进行一系列实验,在驱动图像的模型中进行反事实解释,我们用一个方法来解释。我们用哪种方法可以用来解释。