Deep neural networks (DNNs) are the workhorses of deep learning, which constitutes the state of the art in numerous application domains. However, DNN-based decision rules are notoriously prone to poor generalization, i.e., may prove inadequate on inputs not encountered during training. This limitation poses a significant obstacle to employing deep learning for mission-critical tasks, and also in real-world environments that exhibit high variability. We propose a novel, verification-driven methodology for identifying DNN-based decision rules that generalize well to new input domains. Our approach quantifies generalization to an input domain by the extent to which decisions reached by independently trained DNNs are in agreement for inputs in this domain. We show how, by harnessing the power of DNN verification, our approach can be efficiently and effectively realized. We evaluate our verification-based approach on three deep reinforcement learning (DRL) benchmarks, including a system for real-world Internet congestion control. Our results establish the usefulness of our approach, and, in particular, its superiority over gradient-based methods. More broadly, our work puts forth a novel objective for formal verification, with the potential for mitigating the risks associated with deploying DNN-based systems in the wild.
翻译:深心神经网络(DNN)是深层学习的工马,是许多应用领域的先进水平。然而,基于DNN的决定规则臭名昭著,容易被普遍化,也就是说,培训期间没有遇到的投入可能证明不足。这一限制严重阻碍了对任务关键任务以及具有高度差异的现实世界环境中的深入学习。我们提出了一种创新的、由核查驱动的方法,用于确定基于DNN的决定规则,该方法能够将信息广泛化为新的输入领域。我们的方法根据独立培训的DNNN所达成的决定对该领域投入达成一致的程度量化。我们通过利用DNN核查的力量,表明我们的方法如何能够高效率和有效地实现。我们评估了我们基于三个深度强化学习基准的核查方法,包括现实世界互联网拥堵控制系统。我们的结果确定了我们方法的有用性,特别是它优于基于梯度的方法。更广泛地说,我们的工作提出了正式核查的新目标,即通过利用DNN核查的力量,在部署DN系统时有可能减少与野生相关的风险。