Deep Reinforcement Learning (DRL) has produced great achievements since it was proposed, including the possibility of processing raw vision input data. However, training an agent to perform tasks based on image feedback remains a challenge. It requires the processing of large amounts of data from high-dimensional observation spaces, frame by frame, and the agent's actions are computed according to deep neural network policies, end-to-end. Image pre-processing is an effective way of reducing these high dimensional spaces, eliminating unnecessary information present in the scene, supporting the extraction of features and their representations in the agent's neural network. Modern video-games are examples of this type of challenge for DRL algorithms because of their visual complexity. In this paper, we propose a low-dimensional observation filter that allows a deep Q-network agent to successfully play in a visually complex and modern video-game, called Neon Drive.
翻译:自提出以来,深强化学习(DRL)取得了巨大成就,包括处理原始视觉输入数据的可能性。然而,培训一名代理人员执行基于图像反馈的任务仍是一项挑战。这需要处理来自高维观测空间的大量数据、框架框架和代理人员的行动根据深神经网络政策、端到端计算。图像预处理是减少这些高维空间、消除现场存在的不必要信息、支持提取特征及其在代理人员神经网络中的体现的有效方法。现代视频游戏是DRL算法因其视觉复杂性而面临这类挑战的例子。在本文中,我们提议了一个低维观测过滤器,使深Q网络代理能够在视觉复杂和现代视频游戏中成功播放,称为Neon驱动器。