Deep reinforcement learning (DRL) agents are trained through trial-and-error interactions with the environment. This leads to a long training time for dense neural networks to achieve good performance. Hence, prohibitive computation and memory resources are consumed. Recently, learning efficient DRL agents has received increasing attention. Yet, current methods focus on accelerating inference time. In this paper, we introduce for the first time a dynamic sparse training approach for deep reinforcement learning to accelerate the training process. The proposed approach trains a sparse neural network from scratch and dynamically adapts its topology to the changing data distribution during training. Experiments on continuous control tasks show that our dynamic sparse agents achieve higher performance than the equivalent dense methods, reduce the parameter count and floating-point operations (FLOPs) by 50%, and have a faster learning speed that enables reaching the performance of dense agents with 40-50% reduction in the training steps.
翻译:深层强化学习( DRL) 代理机构通过试验和与环境的试镜互动来培训深强化学习( DRL) 代理机构。 这导致密集神经网络获得良好性能的长时间培训时间。 因此, 计算和记忆资源耗尽了。 最近, 学习高效的 DRL 代理机构受到越来越多的关注。 然而, 目前的方法侧重于加速推论时间 。 在本文中, 我们首次引入了动态稀疏的深层强化学习方法, 以加快培训进程。 拟议的方法从零开始培养一个稀疏的神经网络, 并动态地调整其地形以适应培训期间不断变化的数据分布。 持续控制任务的实验显示, 我们动态稀有代理机构的表现高于同等密集的方法, 将参数计数和浮点操作( FLOPs) 减少50%, 并具有更快的学习速度, 使密度代理机构的表现在培训步骤中减少40- 50% 。