In safety-critical deep learning applications robustness measurement is a vital pre-deployment phase. However, existing robustness verification methods are not sufficiently practical for deploying machine learning systems in the real world. On the one hand, these methods attempt to claim that no perturbations can ``fool'' deep neural networks (DNNs), which may be too stringent in practice. On the other hand, existing works rigorously consider $L_p$ bounded additive perturbations on the pixel space, although perturbations, such as colour shifting and geometric transformations, are more practically and frequently occurring in the real world. Thus, from the practical standpoint, we present a novel and general {\it probabilistic robustness assessment method} (PRoA) based on the adaptive concentration, and it can measure the robustness of deep learning models against functional perturbations. PRoA can provide statistical guarantees on the probabilistic robustness of a model, \textit{i.e.}, the probability of failure encountered by the trained model after deployment. Our experiments demonstrate the effectiveness and flexibility of PRoA in terms of evaluating the probabilistic robustness against a broad range of functional perturbations, and PRoA can scale well to various large-scale deep neural networks compared to existing state-of-the-art baselines. For the purpose of reproducibility, we release our tool on GitHub: \url{ https://github.com/TrustAI/PRoA}.
翻译:在安全的深层学习应用中,稳健度测量是一个至关重要的部署前阶段。然而,现有的稳健度核实方法对于在现实世界中部署机器学习系统并不十分实用。 一方面,这些方法试图声称,不扰动“ fool” 深神经网络(DNNs),实际上可能过于严格。 另一方面,现有的工程严格考虑在像素空间上使用$_p$ 受约束的添加扰动,尽管色彩变化和几何转换等扰动在现实世界中发生得更为实际和频繁。 因此,从实际的角度来看,我们提出了一个基于适应集中的新和一般的实能性强力评估方法}(PRoA),可以测量深层学习模型相对于功能扰动的稳健性。 PROA可以提供统计保证模型的稳健性强性,用于我们所培训的模型在部署后遇到的失败概率。 我们的实验展示了PROA的实效和灵活性,用来对当前大规模运行性网络进行对比。