Transfer Learning (TL) is an efficient machine learning paradigm that allows overcoming some of the hurdles that characterize the successful training of deep neural networks, ranging from long training times to the needs of large datasets. While exploiting TL is a well established and successful training practice in Supervised Learning (SL), its applicability in Deep Reinforcement Learning (DRL) is rarer. In this paper, we study the level of transferability of three different variants of Deep-Q Networks on popular DRL benchmarks as well as on a set of novel, carefully designed control tasks. Our results show that transferring neural networks in a DRL context can be particularly challenging and is a process which in most cases results in negative transfer. In the attempt of understanding why Deep-Q Networks transfer so poorly, we gain novel insights into the training dynamics that characterizes this family of algorithms.
翻译:转移学习(TL)是一种高效的机器学习模式,它能够克服深神经网络成功培训所特有的一些障碍,从培训时间长到大型数据集的需要等。在监督学习(SL)中,利用TL是一种既定的成功培训做法,但在深强化学习(DRL)中,其适用性则更为罕见。在本文中,我们研究了三种不同的深Q网络在流行的DRL基准以及一套新颖、精心设计的控制任务方面的可转移性。我们的结果显示,在DRL背景下转移神经网络可能特别具有挑战性,而且在大多数情况下,这是一个导致负转移的过程。为了理解深Q网络转移如此差的原因,我们从培训动态中获得了新的洞察力。