Transfer in Reinforcement Learning aims to improve learning performance on target tasks using knowledge from experienced source tasks. Successor Representations (SR) and their extension Successor Features (SF) are prominent transfer mechanisms in domains where reward functions change between tasks. They reevaluate the expected return of previously learned policies in a new target task to transfer their knowledge. The SF framework extended SR by linearly decomposing rewards into successor features and a reward weight vector allowing their application in high-dimensional tasks. But this came with the cost of having a linear relationship between reward functions and successor features, limiting its application to such tasks. We propose a novel formulation of SR based on learning the cumulative discounted probability of successor features, called Successor Feature Representations (SFR). Crucially, SFR allows to reevaluate the expected return of policies for general reward functions. We introduce different SFR variations, prove its convergence, and provide a guarantee on its transfer performance. Experimental evaluations based on SFR with function approximation demonstrate its advantage over SF not only for general reward functions but also in the case of linearly decomposable reward functions.
翻译:强化学习转让的目的是利用来自经验丰富的来源任务的知识,提高目标任务的学习成绩; 继任代表(SR)及其扩展继承地物(SF)是不同任务之间奖励职能发生变化的领域的主要转移机制; 重新评估在新的目标任务中以前学到的政策的预期回报,以转移知识; 强化学习框架通过将奖励线性地分解成继承特征和奖励重量矢量将其应用于高层次任务而扩大了SR的范围; 但是,这是在奖励职能与后续特征之间建立线性关系的成本,将其应用限制在这类任务上; 我们建议以学习累积的折扣后继特征概率为基础,即 " 继承特征代表 " (SFR)。 关键是,SFR允许重新评价一般奖励职能的预期回报。 我们采用不同的SFR变式,证明其趋同,并保障其转移业绩。 基于SFR和功能近似的实验性评价表明,它不仅在一般奖励职能上,而且在可线性解奖项职能上比SFF的优势。