Control certificates based on barrier functions have been a powerful tool to generate probably safe control policies for dynamical systems. However, existing methods based on barrier certificates are normally for white-box systems with differentiable dynamics, which makes them inapplicable to many practical applications where the system is a black-box and cannot be accurately modeled. On the other side, model-free reinforcement learning (RL) methods for black-box systems suffer from lack of safety guarantees and low sampling efficiency. In this paper, we propose a novel method that can learn safe control policies and barrier certificates for black-box dynamical systems, without requiring for an accurate system model. Our method re-designs the loss function to back-propagate gradient to the control policy even when the black-box dynamical system is non-differentiable, and we show that the safety certificates hold on the black-box system. Empirical results in simulation show that our method can significantly improve the performance of the learned policies by achieving nearly 100% safety and goal reaching rates using much fewer training samples, compared to state-of-the-art black-box safe control methods. Our learned agents can also generalize to unseen scenarios while keeping the original performance. The source code can be found at https://github.com/Zengyi-Qin/bcbf.
翻译:以屏障功能为基础的控制证书是产生动态系统可能安全控制政策的有力工具。然而,基于屏障证书的现有方法通常是对具有不同动态的白箱系统采用的,这使得这些方法不适用于许多实际应用,而该系统是黑箱,无法精确建模。另一方面,黑箱系统的无型强化学习(RL)方法缺乏安全保障和低取样效率。在本文中,我们提出了一个新方法,可以学习黑箱动态系统的安全控制政策和障碍证书,而不需要精确的系统模型。我们的方法是将损失功能重新设计成向控制政策后方推进梯度,即使黑箱动态系统是无差异的,而且我们表明,安全证书仍保留在黑箱系统上。模拟实验结果显示,我们的方法可以大大改进学习政策的绩效,即实现近100%的安全和目标达标率,而不必使用精确的系统模型。我们所学的代理商也可以在保留原始源码/原始源码的同时,在原始源/原始源/原始码中找到。我们所学的代理人也可以将原始码/Z。