In the context of MDPs with high-dimensional states, reinforcement learning can achieve better results when using a compressed, low-dimensional representation of the original input space. A variety of learning objectives have therefore been used to learn useful representations. However, these representations usually lack interpretability of the different features. We propose a representation learning algorithm that is able to disentangle latent features into a controllable and an uncontrollable part. The resulting representations are easily interpretable and can be used for learning and planning efficiently by leveraging the specific properties of the two parts. To highlight the benefits of the approach, the disentangling properties of the algorithm are illustrated in three different environments.
翻译:在具有高维状态的 MDP 背景下,如果使用原始输入空间的压缩、低维表示法,强化学习可以取得更好的结果。因此,利用各种学习目标来学习有用的表示法。然而,这些表示法通常缺乏对不同特征的解释性。我们建议一种代表学习算法,能够将潜在特征分解为可控和不可控制的部分。由此产生的表示法很容易解释,并且能够通过利用这两个部分的具体特性有效地用于学习和规划。为了突出该方法的好处,在三个不同环境中说明了算法的分解特性。