Within a robotic context, we merge the techniques of passivity-based control (PBC) and reinforcement learning (RL) with the goal of eliminating some of their reciprocal weaknesses, as well as inducing novel promising features in the resulting framework. We frame our contribution in a scenario where PBC is implemented by means of virtual energy tanks, a control technique developed to achieve closed-loop passivity for any arbitrary control input. Albeit the latter result is heavily used, we discuss why its practical application at its current stage remains rather limited, which makes contact with the highly debated claim that passivity-based techniques are associated with a loss of performance. The use of RL allows us to learn a control policy that can be passivized using the energy tank architecture, combining the versatility of learning approaches and the system theoretic properties which can be inferred due to the energy tanks. Simulations show the validity of the approach, as well as novel interesting research directions in energy-aware robotics.
翻译:机器人中学习策略的被动化和使用虚拟能量箱学习被动策略
在机器人环境中,我们将基于被动控制(PBC)和强化学习(RL)的技术相融合,旨在消除相互的弱点并在所得到的框架中引入新的有前途的特性。我们将我们的贡献放在虚拟能量箱实现的PBC框架中,后者是一种开发出来实现对任意控制输入的闭环被动性的控制技术。虽然后者的结果被广泛运用,但我们讨论了它目前的实际应用相对受限,这使得我们关注经常被争议的声明:基于被动性的技术与性能损失相关。使用RL允许我们学习一个可以通过能量箱架构来实现被动化的控制策略,结合了学习方法的灵活性和由于能量箱可以推断得到的系统理论特性。仿真展示了该方法的有效性,以及在面向能源感知的机器人领域中的有趣新研究方向。