Rapid non-verbal communication of task-based stimuli is a challenge in human-machine teaming, particularly in closed-loop interactions such as driving. To achieve this, we must understand the representations of information for both the human and machine, and determine a basis for bridging these representations. Techniques of explainable artificial intelligence (XAI) such as layer-wise relevance propagation (LRP) provide visual heatmap explanations for high-dimensional machine learning techniques such as deep neural networks. On the side of human cognition, visual attention is driven by the bottom-up and top-down processing of sensory input related to the current task. Since both XAI and human cognition should focus on task-related stimuli, there may be overlaps between their representations of visual attention, potentially providing a means of nonverbal communication between the human and machine. In this work, we examine the correlations between LRP heatmap explanations of a neural network trained to predict driving behavior and eye gaze heatmaps of human drivers. The analysis is used to determine the feasibility of using such a technique for enhancing driving performance. We find that LRP heatmaps show increasing levels of similarity with eye gaze according to the task specificity of the neural network. We then propose how these findings may assist humans by visually directing attention towards relevant areas. To our knowledge, our work provides the first known analysis of LRP and eye gaze for driving tasks.
翻译:以快速非语言方式交流基于任务刺激的刺激性技术是人类机器团队的一项挑战,特别是在驾驶等闭路互动中。 为了实现这一目标,我们必须理解人类和机器的信息表达方式,并确定弥补这些表达方式的基础。 可解释的人工智能(XAI)技术,例如分层关联性传播(LRP)为高维机器学习技术,如深神经网络提供了视觉热映图解释。在人类认知方面,视觉关注是由与当前任务有关的感官输入的自下而下和自上而下的处理过程驱动的。由于XAI和人类认知应侧重于与任务相关的刺激性,因此在视觉关注的表达方式之间可能存在重叠,有可能提供人类和机器之间非口头沟通的手段。在这项工作中,我们研究了为预测驱动行为而训练的神经网络的热映图解释关系。我们用这种技术来提高当前任务中与直视直视直观的驱动力分析的可能性。我们发现,LRP热映像系统可以提供与我们已知的直观分析任务相关的研究区域。