While Feedforward Neural Networks (FNNs) have achieved remarkable success in various tasks, they are vulnerable to adversarial examples. Several techniques have been developed to verify the adversarial robustness of FNNs, but most of them focus on robustness verification against the local perturbation neighborhood of a single data point. There is still a large research gap in global robustness analysis. The global-robustness verifiable framework DeepGlobal has been proposed to identify \textit{all} possible Adversarial Dangerous Regions (ADRs) of FNNs, not limited to data samples in a test set. In this paper, we propose a complete specification and implementation of DeepGlobal utilizing the SMT solver Z3 for more explicit definition, and propose several improvements to DeepGlobal for more efficient verification. To evaluate the effectiveness of our implementation and improvements, we conduct extensive experiments on a set of benchmark datasets. Visualization of our experiment results shows the validity and effectiveness of the approach.
翻译:尽管前馈神经网络在各种任务中取得了显着的成功,但是它们容易受到对抗性样本的攻击。已经开发了几种技术来验证FNN的鲁棒性,但大多数技术都集中在单个数据点的局部扰动邻域的鲁棒验证上。全局鲁棒性分析仍存在较大的研究空白。已经提出了 DeepGlobal 全局鲁棒性可验证框架,用于识别FNN的\textit{所有}可能的对抗性危险区域(ADRs),不限于测试集中的数据样本。在本文中,我们提出了DeepGlobal的完整规范和实现,利用SMT求解器Z3进行更明确的定义,并提出了一些改进措施,以实现更高效的验证。为了评估我们的实现和改进的有效性,我们对一组基准数据集进行了广泛的实验。我们的实验结果可视化表明了该方法的有效性和有效性。