Deep neural networks (DNNs) are a state-of-the-art technology, capable of outstanding performance in many key tasks. However, it is challenging to integrate DNNs into safety-critical systems, such as those in the aerospace or automotive domains, due to the risk of adversarial inputs: slightly perturbed inputs that can cause the DNN to make grievous mistakes. Adversarial inputs have been shown to plague even modern DNNs; and so the risks they pose must be measured and mitigated to allow the safe deployment of DNNs in safety-critical systems. Here, we present a novel and scalable tool called gRoMA, which uses a statistical approach for formally measuring the global categorial robustness of a DNN - i.e., the probability of randomly encountering an adversarial input for a specific output category. Our tool operates on pre-trained, black-box classification DNNs. It randomly generates input samples that belong to an output category of interest, measures the DNN's susceptibility to adversarial inputs around these inputs, and then aggregates the results to infer the overall global robustness of the DNN up to some small bounded error. For evaluation purposes, we used gRoMA to measure the global robustness of the widespread Densenet DNN model over the CIFAR10 dataset and our results exposed significant gaps in the robustness of the different output categories. This experiment demonstrates the scalability of the new approach and showcases its potential for allowing DNNs to be deployed within critical systems of interest.
翻译:深心神经网络(DNN)是一种最先进的技术,能够在许多关键任务中取得杰出的成绩。然而,由于对抗性投入的风险,将DNN纳入安全关键系统,例如航空航天或汽车域的安全关键系统是一项艰巨的任务:略有扰动的投入可能会使DNN产生严重错误。对立投入已经显示会困扰甚至现代DNN;因此,它们构成的风险必须加以测量和减轻,以便能够安全地在安全关键系统中部署DNN。在这里,我们提出了一个叫作GRoMA的新而可扩展的工具,它使用统计方法正式测量DNN - 即汽车域的全局性坚固性,即随机遇到一个特定产出类别的对抗性投入的可能性。我们的工具运行于预先训练的黑箱分类 DNNN。它随机生成了属于某一产出类别的输入样本,测量了DNNN在这些投入方面的潜在潜在潜力,然后将结果汇总成一个称为GROMA的临界性模型,显示DNFAR在高度数据中所使用的总的坚固性。