Predictions made by deep neural networks were shown to be highly sensitive to small changes made in the input space where such maliciously crafted data points containing small perturbations are being referred to as adversarial examples. On the other hand, recent research suggests that the same networks can also be extremely insensitive to changes of large magnitude, where predictions of two largely different data points can be mapped to approximately the same output. In such cases, features of two data points are said to approximately collide, thus leading to the largely similar predictions. Our results improve and extend the work of Li et al.(2019), laying out theoretical grounds for the data points that have colluding features from the perspective of weights of neural networks, revealing that neural networks not only suffer from features that approximately collide but also suffer from features that exactly collide. We identify the necessary conditions for the existence of such scenarios, hereby investigating a large number of DNNs that have been used to solve various computer vision problems. Furthermore, we propose the Null-space search, a numerical approach that does not rely on heuristics, to create data points with colliding features for any input and for any task, including, but not limited to, classification, localization, and segmentation.
翻译:深神经网络的预测显示,深神经网络对输入空间的微小变化非常敏感,其中含有小扰动的恶意编造的数据点被称作对抗性实例。另一方面,最近的研究表明,同样的网络也可能对巨大变化极不敏感,对两个大相径庭的数据点的预测可以绘制出大致相同的结果。在这种情况下,两个数据点的特征据说大致相撞,从而导致大致相似的预测。我们的结果改进并扩展了Li等人(2019年)的工作,从神经网络的权重角度为具有相互勾结特征的数据点提供了理论依据,表明神经网络不仅具有大约相撞的特征,而且还有完全相撞的特征。我们确定了存在这种情景的必要条件,在此情况下,我们调查大量用于解决各种计算机视觉问题的DNN(DN)点的特征。此外,我们提议使用Null空间搜索,一种不依赖神经学的数值方法,但不依赖超理论,而是创造数据点,包括任何输入和分类的局部分段,包括任何输入和分级的分段。