The widespread use of deep neural networks has achieved substantial success in many tasks. However, there still exists a huge gap between the operating mechanism of deep learning models and human-understandable decision making, so that humans cannot fully trust the predictions made by these models. To date, little work has been done on how to align the behaviors of deep learning models with human perception in order to train a human-understandable model. To fill this gap, we propose a new framework to train a deep neural network by incorporating the prior of human perception into the model learning process. Our proposed model mimics the process of perceiving conceptual parts from images and assessing their relative contributions towards the final recognition. The effectiveness of our proposed model is evaluated on two classical visual recognition tasks. The experimental results and analysis confirm our model is able to provide interpretable explanations for its predictions, but also maintain competitive recognition accuracy.
翻译:然而,深层神经网络的广泛使用在许多任务中取得了巨大成功。然而,深层学习模型的运作机制与人类可以理解的决策之间仍然存在着巨大差距,因此人类无法完全相信这些模型所作的预测。迄今为止,在如何使深层学习模型的行为与人类感知相一致以便培养一个人类能理解的模式方面,没有做多少工作。为了填补这一差距,我们提议一个新的框架,通过将人类先前的认知纳入模型学习过程来培训深层神经网络。我们提议的模型模仿了从图像中发现概念部分的过程,并评估其对最终认知的相对贡献。我们提议的模型的有效性是通过两种古典视觉识别任务进行评估的。实验结果和分析证实我们的模型能够为其预测提供可解释的解释,但也保持了竞争性认知的准确性。