Deep neural networks (DNNs) have gained significant popularity in recent years, becoming the state of the art in a variety of domains. In particular, deep reinforcement learning (DRL) has recently been employed to train DNNs that act as control policies for various types of real-world systems. In this work, we present the whiRL 2.0 tool, which implements a new approach for verifying complex properties of interest for such DRL systems. To demonstrate the benefits of whiRL 2.0, we apply it to case studies from the communication networks domain that have recently been used to motivate formal verification of DRL systems, and which exhibit characteristics that are conducive for scalable verification. We propose techniques for performing k-induction and automated invariant inference on such systems, and use these techniques for proving safety and liveness properties of interest that were previously impossible to verify due to the scalability barriers of prior approaches. Furthermore, we show how our proposed techniques provide insights into the inner workings and the generalizability of DRL systems. whiRL 2.0 is publicly available online.
翻译:近年来,深神经网络(DNNs)已获得显著的欢迎,成为各个领域的先进技术,特别是,最近利用深强化学习(DRL)来培训作为各种现实世界系统控制政策的DNS。在这项工作中,我们介绍了THERL 2.0工具,该工具采用新的方法来核查这些DRL系统感兴趣的复杂特性。为了展示WHERL 2.0的好处,我们将其应用于通信网络域域的案例研究,这些案例研究最近被用来激励对DRL系统进行正式核查,并展示了有利于可扩展核查的特性。我们提出了对此类系统进行 k 上传和自动变异推论的技术,并使用这些技术来证明先前由于先前方法的可扩展性障碍而无法核实的安全和活性特性。此外,我们展示了我们所提议的技术如何为DRL系统的内部工作和可普及性提供洞察力。THER 2.0可公开在线查阅。