Implicit neural networks are a general class of learning models that replace the layers in traditional feedforward models with implicit algebraic equations. Compared to traditional learning models, implicit networks offer competitive performance and reduced memory consumption. However, they can remain brittle with respect to input adversarial perturbations. This paper proposes a theoretical and computational framework for robustness verification of implicit neural networks; our framework blends together mixed monotone systems theory and contraction theory. First, given an implicit neural network, we introduce a related embedded network and show that, given an $\ell_\infty$-norm box constraint on the input, the embedded network provides an $\ell_\infty$-norm box overapproximation for the output of the given network. Second, using $\ell_{\infty}$-matrix measures, we propose sufficient conditions for well-posedness of both the original and embedded system and design an iterative algorithm to compute the $\ell_{\infty}$-norm box robustness margins for reachability and classification problems. Third, of independent value, we propose a novel relative classifier variable that leads to tighter bounds on the certified adversarial robustness in classification problems. Finally, we perform numerical simulations on a Non-Euclidean Monotone Operator Network (NEMON) trained on the MNIST dataset. In these simulations, we compare the accuracy and run time of our mixed monotone contractive approach with the existing robustness verification approaches in the literature for estimating the certified adversarial robustness.
翻译:隐性神经网络是一种普通学习模式, 以隐含的代数方程式取代传统进料模型的层层。 与传统学习模式相比, 隐性网络提供竞争性性能和减少内存消耗。 但是, 与输入对称扰动相比, 隐性神经网络仍然会变得很弱。 本文提出一个理论和计算框架, 用于对隐性神经网络进行稳健性核查; 我们的框架将单质系统理论和收缩理论混合在一起。 首先, 鉴于隐含的神经网络, 我们引入了一个相关的内嵌网络, 并显示, 与传统的学习模式相比, 隐性网络提供了一种 $\ incenty$- inty$- 诺性框限制, 内嵌性网络输出有一个 $\ell inty- entrmormexcity 框内, 内含经认证的内置性内置的内置性内置值, 内置的内置性内置内置内置内置内置内置内置内置的内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置的内置内置内置内置内置的内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置的内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内置内