Although transfer learning is considered to be a milestone in deep reinforcement learning, the mechanisms behind it are still poorly understood. In particular, predicting if knowledge can be transferred between two given tasks is still an unresolved problem. In this work, we explore the use of network distillation as a feature extraction method to better understand the context in which transfer can occur. Notably, we show that distillation does not prevent knowledge transfer, including when transferring from multiple tasks to a new one, and we compare these results with transfer without prior distillation. We focus our work on the Atari benchmark due to the variability between different games, but also to their similarities in terms of visual features.
翻译:虽然转移学习被认为是深层强化学习中的一个里程碑,但其背后的机制仍然不易理解,特别是预测知识能否在两个特定任务之间转移仍然是一个尚未解决的问题。在这项工作中,我们探索使用网络蒸馏法作为特征提取方法,以更好地了解转让的背景。值得注意的是,我们表明,蒸馏并不妨碍知识转让,包括从多重任务转移到新的任务,我们将这些结果与没有事先蒸馏的转让进行比较。我们把工作重点放在阿塔里基准上,因为不同游戏之间互不相同,而且视像特征相似。