In high-dimensional state spaces, the usefulness of Reinforcement Learning (RL) is limited by the problem of exploration. This issue has been addressed using potential-based reward shaping (PB-RS) previously. In the present work, we introduce Final-Volume-Preserving Reward Shaping (FV-RS). FV-RS relaxes the strict optimality guarantees of PB-RS to a guarantee of preserved long-term behavior. Being less restrictive, FV-RS allows for reward shaping functions that are even better suited for improving the sample efficiency of RL algorithms. In particular, we consider settings in which the agent has access to an approximate plan. Here, we use examples of simulated robotic manipulation tasks to demonstrate that plan-based FV-RS can indeed significantly improve the sample efficiency of RL over plan-based PB-RS.
翻译:在高地国家空间,强化学习的效用受到勘探问题的限制,这个问题已经利用以前基于潜在奖励的形状(PB-RS)来解决,在目前的工作中,我们采用了最终-Volume-Preserve Reward 形状(FV-RS),FV-RS放松了PB-RS的严格最佳保证,以保障保存的长期行为。FV-RS限制较少,允许奖励甚至更适合提高RL算法抽样效率的塑造功能。我们特别考虑到该代理商可以获得近似计划的环境。在这里,我们使用模拟机器人操纵任务的例子来证明,基于计划的FV-RS确实能够大大提高RL超过基于计划的PB-RS的样本效率。