Neural Networks, being susceptible to adversarial attacks, should face a strict level of scrutiny before being deployed in critical or adversarial applications. This paper uses ideas from Chaos Theory to explain, analyze, and quantify the degree to which Neural Networks are susceptible to or robust against adversarial attacks. Our results show that susceptibility to attack grows significantly with the depth of the model, which has significant safety implications for the design of Neural Networks for production environments. We also demonstrate how to quickly and easily approximate the certified robustness radii for extremely large models, which until now has been computationally infeasible to calculate directly, as well as show a clear relationship between our new susceptibility metric and post-attack accuracy.
翻译:神经网络容易遭到对抗性攻击,在被部署到关键或对抗性应用程序之前,应受到严格审查。本文使用混乱理论的观点来解释、分析和量化神经网络在对抗性攻击时的易遭受程度或强力。我们的结果显示,随着模型的深度,攻击的易感性将大大增加,这对神经网络在生产环境中的设计具有重大安全影响。我们还表明如何迅速和容易地将经核证的强力射线对极大型模型进行近似,而迄今为止,这些模型一直无法直接计算,并表明我们新的易感度度度度和攻击后精确度之间的明确关系。