Reachability analysis is a promising technique to automatically prove or disprove the reliability and safety of AI-empowered software systems that are developed by using Deep Reinforcement Learning (DRL). Existing approaches suffer however from limited scalability and large overestimation as they must over-approximate the complex and almost inexplicable system components, namely deep neural networks (DNNs). In this paper we propose a novel, tight and scalable reachability analysis approach for DRL systems. By training on abstract states, our approach treats the embedded DNNs as black boxes to avoid the over-approximation for neural networks in computing reachable sets. To tackle the state explosion problem inherent to abstraction-based approaches, we devise a novel adjacent interval aggregation algorithm which balances the growth of abstract states and the overestimation caused by the abstraction. We implement a tool, called BBReach, and assess it on an extensive benchmark of control systems to demonstrate its tightness, scalability, and efficiency.
翻译:可实现性分析是自动证明或否定使用深强化学习开发的AI型软件系统的可靠性和安全性的一个大有希望的技术。但现有方法的可扩缩性和高估度有限,因为它们必须过分接近复杂和几乎无法解释的系统组件,即深神经网络。在本文件中,我们为DRL系统提出了一个新颖、紧凑和可扩缩的可达性分析方法。通过对抽象国家的培训,我们的方法将嵌入的DNN作为黑盒,以避免神经网络在计算可达数据集中的过度接近。为了解决抽象方法所固有的国家爆炸问题,我们设计了一个新的相邻间间集算法,以平衡抽象状态的增长和抽象造成的过高估计。我们实施了一个工具,称为BBriacheach, 并评估其控制系统的广泛基准,以证明其紧凑性、可伸缩性和效率。