Reinforcement learning (RL) has shown a promising performance in learning optimal policies for a variety of sequential decision-making tasks. However, in many real-world RL problems, besides optimizing the main objectives, the agent is expected to satisfy a certain level of safety (e.g., avoiding collisions in autonomous driving). While RL problems are commonly formalized as Markov decision processes (MDPs), safety constraints are incorporated via constrained Markov decision processes (CMDPs). Although recent advances in safe RL have enabled learning safe policies in CMDPs, these safety requirements should be satisfied during both training and in the deployment process. Furthermore, it is shown that in memory-based and partially observable environments, these methods fail to maintain safety over unseen out-of-distribution observations. To address these limitations, we propose a Lyapunov-based uncertainty-aware safe RL model. The introduced model adopts a Lyapunov function that converts trajectory-based constraints to a set of local linear constraints. Furthermore, to ensure the safety of the agent in highly uncertain environments, an uncertainty quantification method is developed that enables identifying risk-averse actions through estimating the probability of constraint violations. Moreover, a Transformers model is integrated to provide the agent with memory to process long time horizons of information via the self-attention mechanism. The proposed model is evaluated in grid-world navigation tasks where safety is defined as avoiding static and dynamic obstacles in fully and partially observable environments. The results of these experiments show a significant improvement in the performance of the agent both in achieving optimality and satisfying safety constraints.
翻译:强化学习(RL)在为各种顺序决策任务学习最佳政策方面表现出了良好的业绩,然而,在许多现实世界的RL问题中,除了优化主要目标外,还有望满足一定程度的安全(例如避免自动驾驶过程中的碰撞),尽管RL问题通常通过Markov决策过程(MDPs)而正式化,但安全限制通过限制的Markov决策过程(MCDs)纳入。虽然安全RL最近的进展使得CMDP能够学习安全政策,但在培训和部署过程中,这些安全要求都应得到满足。此外,在基于记忆和部分观察的环境中,这些方法无法维持某种程度的安全(例如避免在自主驾驶过程中的碰撞)。虽然RLyapunov问题通常作为Markov决策过程(MDPs)正式化,但采用Lyapunov的静态功能,将基于轨迹的限制转换为一套局部线性限制。此外,为确保代理人在高度不确定的环境中的安全,应当制定一种不确定的量化方法,以便能够在基于记忆和部分观察的环境中识别风险的行动。为了解决在长期的轨道上发生的障碍,拟议中,在通过精确的精确的核查过程中,从保证实现一种精确的精确的精确的精确性规则的工能提供一种完整的评价。