Artificial intelligence (AI) systems power the world we live in. Deep neural networks (DNNs) are able to solve tasks in an ever-expanding landscape of scenarios, but our eagerness to apply these powerful models leads us to focus on their performance and deprioritises our ability to understand them. Current research in the field of explainable AI tries to bridge this gap by developing various perturbation or gradient-based explanation techniques. For images, these techniques fail to fully capture and convey the semantic information needed to elucidate why the model makes the predictions it does. In this work, we develop a new form of explanation that is radically different in nature from current explanation methods, such as Grad-CAM. Perception visualization provides a visual representation of what the DNN perceives in the input image by depicting what visual patterns the latent representation corresponds to. Visualizations are obtained through a reconstruction model that inverts the encoded features, such that the parameters and predictions of the original models are not modified. Results of our user study demonstrate that humans can better understand and predict the system's decisions when perception visualizations are available, thus easing the debugging and deployment of deep models as trusted systems.
翻译:深神经网络(DNNS)能够在一个不断扩大的情景环境中解决任务,但我们热切地应用这些强大的模型导致我们专注于其性能,并消除了理解这些模型的能力。 目前在可解释的AI领域的研究试图通过开发各种扰动或梯度解释技术来弥合这一差距。 对于图像来说,这些技术无法充分捕捉和传递解释模型为何作出预测所需的语义信息。 在这项工作中,我们开发了一种新的解释形式,其性质与当前解释方法截然不同,如格拉德-卡姆。 视觉化通过描述潜在代表的视觉模式来直观地描述数字图象中的内容。 视觉化是通过一个转换编码特征的重建模型获得的。 原始模型的参数和预测没有被修改。 我们用户研究的结果表明,当人们能够更好地理解和预测系统在可信任的视觉化时,系统将决定作为深层次的定位,因此可以放松。