Within a robotic context, we merge the techniques of passivity-based control (PBC) and reinforcement learning (RL) with the goal of eliminating some of their reciprocal weaknesses, as well as inducing novel promising features in the resulting framework. We frame our contribution in a scenario where PBC is implemented by means of virtual energy tanks, a control technique developed to achieve closed-loop passivity for any arbitrary control input. Albeit the latter result is heavily used, we discuss why its practical application at its current stage remains rather limited, which makes contact with the highly debated claim that passivity-based techniques are associated to a loss of performance. The use of RL allows to learn a control policy which can be passivized using the energy tank architecture, combining the versatility of learning approaches and the system theoretic properties which can be inferred due to the energy tanks. Simulations show the validity of the approach, as well as novel interesting research directions in energy-aware robotics.
翻译:在机器人环境下,我们将被动控制(PBC)和强化学习(RL)技术与消除其一些相互弱点以及由此形成的框架中产生新的有希望的特点的目标结合起来。我们将我们的贡献放在一个假设中,即通过虚拟能源罐(一种为实现任意控制输入而实现闭环被动而开发的控制技术)来实施被动控制(PBC)和强化学习(RL)技术。尽管后者被大量使用,但我们讨论了其在当前阶段的实际应用为何仍然相当有限,这与高度辩论的关于被动技术与性能损失相关联的说法产生了联系。使用RL使得能够学习一种控制政策,而这种控制政策可以使用能源罐结构进行被动化,将学习方法的多功能和因能源罐而可以推断的系统理论性特性结合起来。模拟显示了这种方法的有效性,以及能源智能机器人方面新的令人感兴趣的研究方向。