Reinforcement learning has shown great potential in solving complex tasks when large amounts of data can be generated with little effort. In robotics, one approach to generate training data builds on simulations based on dynamics models derived from first principles. However, for tasks that, for instance, involve complex soft robots, devising such models is substantially more challenging. Being able to train effectively in increasingly complicated scenarios with reinforcement learning enables to take advantage of complex systems such as soft robots. Here, we leverage the imbalance in complexity of the dynamics to learn more sample-efficiently. We (i) abstract the task into distinct components, (ii) off-load the simple dynamics parts into the simulation, and (iii) multiply these virtual parts to generate more data in hindsight. Our new method, Hindsight States (HiS), uses this data and selects the most useful transitions for training. It can be used with an arbitrary off-policy algorithm. We validate our method on several challenging simulated tasks and demonstrate that it improves learning both alone and when combined with an existing hindsight algorithm, Hindsight Experience Replay (HER). Finally, we evaluate HiS on a physical system and show that it boosts performance on a complex table tennis task with a muscular robot. Videos and code of the experiments can be found on webdav.tuebingen.mpg.de/his/.
翻译:强化学习在解决复杂任务方面显示出巨大的潜力:当大量数据能够以很少的努力生成时,大量数据就能产生大量的数据。在机器人中,一种生成培训数据的方法建立在基于根据最初原则产生的动态模型的模拟基础上。然而,对于例如涉及复杂的软机器人的任务,设计这种模型则更具挑战性。能够有效地在日益复杂的情景下进行训练,而强化学习能够利用软机器人等复杂系统。在这里,我们利用动态的不均衡性能来更高效地学习。我们(一) 将这一任务抽象地纳入不同的部件,(二) 将简单的动态部分卸载到模拟中,(三) 将这些虚拟部分乘以在后视中生成更多的数据。我们的新方法,即Hindsight States(HIS),使用这些数据并选择最有用的培训过渡方法。它可以用任意的离政策算法来使用。我们用一些具有挑战性的模拟任务来验证我们的方法,并证明它与现有的后视算法结合, Hindsight Replay (HER) 和(CHER) 增加这些虚拟部分的虚拟部分。最后,我们在物理系统上评估HISS, 并显示一个复合的网格/breabal tabal tablex tad) 。</s>