A domain-theoretic framework is presented for validated robustness analysis of neural networks. First, global robustness of a general class of networks is analyzed. Then, using the fact that Edalat's domain-theoretic L-derivative coincides with Clarke's generalized gradient, the framework is extended for attack-agnostic local robustness analysis. The proposed framework is ideal for designing algorithms which are correct by construction. This claim is exemplified by developing a validated algorithm for estimation of Lipschitz constant of feedforward regressors. The completeness of the algorithm is proved over differentiable networks, and also over general position ReLU networks. Computability results are obtained within the framework of effectively given domains. Using the proposed domain model, differentiable and non-differentiable networks can be analyzed uniformly. The validated algorithm is implemented using arbitrary-precision interval arithmetic, and the results of some experiments are presented. The software implementation is truly validated, as it handles floating-point errors as well.
翻译:为了对神经网络进行经过验证的稳健性分析,介绍了一个域论框架。首先,对一般网络类别的全球稳健性进行了分析。然后,利用Edalat的域-理论L-变性与Clarke的通用梯度相吻合这一事实,扩展了框架,以进行攻击-不可知本地稳健性分析。拟议框架是设计通过构建正确算法的理想框架。这一索赔的例证是,为估算Lipschitz 向后退者进料常数开发了一种经过验证的算法。算法的完整性被证明超过不同的网络,也超过了一般位置 ReLU 网络。在有效给定域的框架内获得了可计算的结果。使用拟议的域模型,可以对不同和无差别的网络进行统一分析。经验证的算法是使用任意的精密间隙算法实施的,一些实验的结果也得到了介绍。软件的实施得到了真正的验证,因为它也处理浮点错误。