The robustness of neural networks is fundamental to the hosting system's reliability and security. Formal verification has been proven to be effective in providing provable robustness guarantees. To improve the verification scalability, over-approximating the non-linear activation functions in neural networks by linear constraints is widely adopted, which transforms the verification problem into an efficiently solvable linear programming problem. As over-approximations inevitably introduce overestimation, many efforts have been dedicated to defining the tightest possible approximations. Recent studies have however showed that the existing so-called tightest approximations are superior to each other. In this paper we identify and report an crucial factor in defining tight approximations, namely the approximation domains of activation functions. We observe that existing approaches only rely on overestimated domains, while the corresponding tight approximation may not necessarily be tight on its actual domain. We propose a novel under-approximation-guided approach, called dual-approximation, to define tight over-approximations and two complementary under-approximation algorithms based on sampling and gradient descent. The overestimated domain guarantees the soundness while the underestimated one guides the tightness. We implement our approach into a tool called DualApp and extensively evaluate it on a comprehensive benchmark of 84 collected and trained neural networks with different architectures. The experimental results show that DualApp outperforms the state-of-the-art approximation-based approaches, with up to 71.22% improvement to the verification result.
翻译:神经网络的稳健性是托管系统的可靠性和安全性的根本。 正式核查已证明在提供可证实的稳健性保证方面是有效的。 为了改进核查的可扩展性,广泛采用线性限制过度使用神经网络的非线性激活功能,从而将核查问题转化为高效溶解线性编程问题。 由于过度使用会不可避免地引入过高估计,许多努力都致力于界定最接近点。 但是,最近的研究表明,现有的所谓最接近点在提供可证实的稳健性保障方面是有效的。 在本文件中,我们确定并报告一个关键因素,用以界定紧凑的近似性,即激活功能的近似性领域。我们观察到,现有方法只依赖高估的域,而相应的紧凑性可能不一定在其实际域内形成紧凑的线性编程问题。我们建议一种新的低压制指导方法,即双相近效制,以界定紧凑的超紧凑度和两次不相近似性准度算法。我们通过抽样和渐渐渐渐变的精确性校准的网络,我们估计了一种超近度的域域域内结果,我们用了一种压式的精确性标标,然后又用一种压式的缩式的校正标标标标, 。我们用一个深地标定的校正的校正的校正的校正的校正的校正的校正的校正的校正的校正的校正的校正的校正的校准结果的校准结果的校准结果的校准结果。