As the third-generation neural networks, Spiking Neural Networks (SNNs) have great potential on neuromorphic hardware because of their high energy-efficiency. However, Deep Spiking Reinforcement Learning (DSRL), i.e., the Reinforcement Learning (RL) based on SNNs, is still in its preliminary stage due to the binary output and the non-differentiable property of the spiking function. To address these issues, we propose a Deep Spiking Q-Network (DSQN) in this paper. Specifically, we propose a directly-trained deep spiking reinforcement learning architecture based on the Leaky Integrate-and-Fire (LIF) neurons and Deep Q-Network (DQN). Then, we adapt a direct spiking learning algorithm for the Deep Spiking Q-Network. We further demonstrate the advantages of using LIF neurons in DSQN theoretically. Comprehensive experiments have been conducted on 17 top-performing Atari games to compare our method with the state-of-the-art conversion method. The experimental results demonstrate the superiority of our method in terms of performance, stability, robustness and energy-efficiency. To the best of our knowledge, our work is the first one to achieve state-of-the-art performance on multiple Atari games with the directly-trained SNN.
翻译:第三代神经网络Spiking Neal Network(SNN)由于高能效,在神经形态硬件方面潜力巨大。然而,深吸强化学习(DSRL),即基于SNNS的强化学习(RL),由于二进制输出和弹出功能的不可区别特性,仍处于初步阶段。为了解决这些问题,我们提议在本文件中建立一个深吸Q-Network(DSQN),具体地说,我们建议基于Leaky Iffil-Fire(LIF)神经元和深Q-Network(DQN)的直接训练深透透的强化学习结构。然后,我们调整了深Spiking Q-Network(RLLL)的直接跳动学习算法。我们进一步展示了在DSQNN理论上使用LIF神经元的优势。我们在17个最优秀的Atari游戏上进行了全面实验,将我们的方法与最先进的转换方法进行比较。 实验结果显示我们最强的Snority-lat the the the preal palalalalalal ex astialal astistrate astistral astistral the the pal ex the the paltistral