Deep reinforcement learning (RL) algorithms are powerful tools for solving visuomotor decision tasks. However, the trained models are often difficult to interpret, because they are represented as end-to-end deep neural networks. In this paper, we shed light on the inner workings of such trained models by analyzing the pixels that they attend to during task execution, and comparing them with the pixels attended to by humans executing the same tasks. To this end, we investigate the following two questions that, to the best of our knowledge, have not been previously studied. 1) How similar are the visual features learned by RL agents and humans when performing the same task? and, 2) How do similarities and differences in these learned features explain RL agents' performance on these tasks? Specifically, we compare the saliency maps of RL agents against visual attention models of human experts when learning to play Atari games. Further, we analyze how hyperparameters of the deep RL algorithm affect the learned features and saliency maps of the trained agents. The insights provided by our results have the potential to inform novel algorithms for the purpose of closing the performance gap between human experts and deep RL agents.
翻译:深度强化学习(RL)算法是解决相对自动决定任务的有力工具。然而,经过训练的模型往往很难解释,因为它们是端至端深神经网络。在本文中,我们通过分析这些经过训练的模型在任务执行期间所关注的像素,并把它们与执行同样任务的人类所关注的像素进行比较,来说明这些模型的内部作用。为此,我们调查了以下两个问题,根据我们的知识,我们没有研究过这两个问题。 1)这些经过训练的模型在从事同样任务时所学到的视觉特征与人类的视觉特征如何相似?和2)这些学到的特征的相似性和差异如何解释RL代理人在这些任务上的绩效?具体地说,我们比较了RL代理人的显性图和人类专家在玩阿塔里游戏时所关注的视觉模型。此外,我们分析了深RL算法的超参数如何影响受过训练的代理人所了解的特征和突出的地图。我们提供的观察结果有可能为缩小人类专家与深RL代理人之间业绩差距的新算法提供新的算法。