Plasticity, the ability of a neural network to quickly change its predictions in response to new information, is essential for the adaptability and robustness of deep reinforcement learning systems. Deep neural networks are known to lose plasticity over the course of training even in relatively simple learning problems, but the mechanisms driving this phenomenon are still poorly understood. This paper conducts a systematic empirical analysis into plasticity loss, with the goal of understanding the phenomenon mechanistically in order to guide the future development of targeted solutions. We find that loss of plasticity is deeply connected to changes in the curvature of the loss landscape, but that it typically occurs in the absence of saturated units or divergent gradient norms. Based on this insight, we identify a number of parameterization and optimization design choices which enable networks to better preserve plasticity over the course of training. We validate the utility of these findings in larger-scale learning problems by applying the best-performing intervention, layer normalization, to a deep RL agent trained on the Arcade Learning Environment.
翻译:神经网络的可塑性,即神经网络根据新信息迅速改变其预测的能力,对于深强化学习系统的适应性和稳健性至关重要。深神经网络在培训过程中,即使在相对简单的学习问题中也已知会丧失可塑性,但造成这一现象的机制仍然不甚为人知。本文件对可塑性损失进行系统的经验分析,目的是机械地理解这一现象,以便指导未来制定有针对性的解决方案。我们发现,可塑性损失与损失景观的曲缩变化密切相关,但通常发生在没有饱和的单元或不同的梯度规范的情况下。我们根据这一观察,确定了一些参数化和优化设计选择,使网络能够在培训过程中更好地保持可塑性。我们通过应用最佳干预、层常规化和在Arcade学习环境中受过训练的深层RL代理物,来验证这些发现在大规模学习问题上的效用。</s>