Control Barrier Functions (CBFs) have been applied to provide safety guarantees for robot navigation. Traditional approaches consider fixed CBFs during navigation and hand-tune the underlying parameters apriori. Such approaches are inefficient and vulnerable to changes in the environment. The goal of this paper is to learn CBFs for multi-robot navigation based on what robots perceive about their environment. In order to guarantee the feasibility of the navigation task, while ensuring robot safety, we pursue a trade-off between conservativeness and aggressiveness in robot behavior by defining dynamic environment-aware CBF constraints. Since the explicit relationship between CBF constraints and navigation performance is challenging to model, we leverage reinforcement learning to learn time-varying CBFs in a model-free manner. We parameterize the CBF policy with graph neural networks (GNNs), and design GNNs that are translation invariant and permutation equivariant, to synthesize decentralized policies that generalize across environments. The proposed approach maintains safety guarantees (due to the underlying CBFs), while optimizing navigation performance (due to the reward-based learning). We perform simulations that compare the proposed approach with fixed CBFs tuned by exhaustive grid-search. The results show that environment-aware CBFs are capable of adapting to robot movements and obstacle changes, yielding improved navigation performance and robust generalization.
翻译:应用了控制屏障功能(CBFs)来为机器人导航提供安全保障。传统方法在航行期间考虑固定的 CBFs,并且将基本参数作为首要的参数。这些方法效率低,易受环境变化的影响。本文件的目的是根据机器人对其环境的看法学习多机器人导航的 CBFs 。为了保证导航任务的可行性,在确保机器人安全的同时,我们通过界定动态环境觉醒的CBF限制,在机器人行为的保守性和侵略性之间实行权衡取舍。由于CBF限制和导航性能之间的明确关系对模型来说具有挑战性,我们利用强化学习来学习时间变化的CBFFs,以无模式的方式学习。我们把CFFS政策与图形神经网络(GNS)进行参数化,并设计能翻译变异和变异的GNNFs,以综合各种环境的分散化政策。拟议办法维持了安全保障(由于基本的CBFFS),同时优化了导航性能(由于有报酬的学习),我们进行了模拟,将CWFS-BS-BS-BS-BS-S-AD-SD-SDRAD-S-SD-S-SD-SD-SD-SD-SDRDRDRDRD-S-S-S-SD-SD-SD-SD-S-S-S-S-S-S-SD-S-S-S-S-SD-SD-SD-SD-SD-SD-SD-SD-SD-SD-SD-SD-SD-S-S-SD-SD-SD-S-S-S-S-S-S-SD-S-S-S-S-S-S-S-S-S-SD-S-S-S-S-S-SD-SD-S-SD-SD-S-S-SD-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-SD-S-S-S-S-S-B-S-S-S-S-</s>