Understanding decisions made by neural networks is key for the deployment of intelligent systems in real world applications. However, the opaque decision making process of these systems is a disadvantage where interpretability is essential. Many feature-based explanation techniques have been introduced over the last few years in the field of machine learning to better understand decisions made by neural networks and have become an important component to verify their reasoning capabilities. However, existing methods do not allow statements to be made about the uncertainty regarding a feature's relevance for the prediction. In this paper, we introduce Monte Carlo Relevance Propagation (MCRP) for feature relevance uncertainty estimation. A simple but powerful method based on Monte Carlo estimation of the feature relevance distribution to compute feature relevance uncertainty scores that allow a deeper understanding of a neural network's perception and reasoning.
翻译:神经网络的决策过程理解对于在实际应用中部署智能系统至关重要。然而,这些系统不透明的决策制定过程在解释性关键的情况下是不利的。为了更好地理解神经网络的决策,过去几年中,机器学习领域已经引入了许多基于特征的解释技术,并已成为验证其推理能力的重要组成部分。然而,现有方法不允许对特征的相关性提供不确定性的陈述。在本文中,我们引入了Monte Carlo Relevance Propagation (MCRP)用于特征相关性不确定性评估。该方法基于Monte Carlo估计特征相关性分布的简单但强大的方法,可以计算特征相关性不确定性分数,从而更深入地理解神经网络的感知和推理机制。