Deep learning (DL) based computer vision (CV) models are generally considered as black boxes due to poor interpretability. This limitation impedes efficient diagnoses or predictions of system failure, thereby precluding the widespread deployment of DLCV models in safety-critical tasks such as autonomous driving. This study is motivated by the need to enhance the interpretability of DL model in autonomous driving and therefore proposes an explainable DL-based framework that generates textual descriptions of the driving environment and makes appropriate decisions based on the generated descriptions. The proposed framework imitates the learning process of human drivers by jointly modeling the visual input (images) and natural language, while using the language to induce the visual attention in the image. The results indicate strong explainability of autonomous driving decisions obtained by focusing on relevant features from visual inputs. Furthermore, the output attention maps enhance the interpretability of the model not only by providing meaningful explanation to the model behavior but also by identifying the weakness of and potential improvement directions for the model.
翻译:深度学习(DL)基于计算机视觉(CV)模型通常被视为黑盒,因为解释能力差,这种局限性妨碍对系统故障的有效诊断或预测,从而无法在诸如自主驾驶等安全关键任务中广泛应用DLV模型,这一研究的动机是需要提高DL模型在自主驾驶中的可解释性,因此提出一个基于DL的可解释框架,生成对驱动环境的文字描述,并根据生成的描述作出适当决定。拟议框架仿照人类驱动者的学习过程,联合制作视觉输入(图像)和自然语言的模型,同时使用该语言吸引图像的视觉关注。结果显示,通过侧重于视觉输入的相关特征而获得的自主驱动决定具有很强的可解释性。此外,产出关注图不仅通过对模型行为提供有意义的解释,而且通过确定模型的弱点和可能的改进方向,提高了模型的可解释性。