As designers of artificial intelligence try to outwit hackers, both sides continue to hone in on AI's inherent vulnerabilities. Designed and trained from certain statistical distributions of data, AI's deep neural networks (DNNs) remain vulnerable to deceptive inputs that violate a DNN's statistical, predictive assumptions. Before being fed into a neural network, however, most existing adversarial examples cannot maintain malicious functionality when applied to an affine transformation. For practical purposes, maintaining that malicious functionality serves as an important measure of the robustness of adversarial attacks. To help DNNs learn to defend themselves more thoroughly against attacks, we propose an affine-invariant adversarial attack, which can consistently produce more robust adversarial examples over affine transformations. For efficiency, we propose to disentangle current affine-transformation strategies from the Euclidean geometry coordinate plane with its geometric translations, rotations and dilations; we reformulate the latter two in polar coordinates. Afterwards, we construct an affine-invariant gradient estimator by convolving the gradient at the original image with derived kernels, which can be integrated with any gradient-based attack methods. Extensive experiments on ImageNet, including some experiments under physical condition, demonstrate that our method can significantly improve the affine invariance of adversarial examples and, as a byproduct, improve the transferability of adversarial examples, compared with alternative state-of-the-art methods.
翻译:人工智能设计者试图超越黑客。 人工智能设计者在试图超越黑客时, 双方都继续用AI 的内在弱点进行磨擦。 根据数据的某些统计分布设计和培训, AI 深神经网络(DNNS) 仍然容易受到违反DNN的统计预测假设的欺骗性投入的伤害。 但是, 在被输入神经网络之前, 大部分现有的对抗性例子无法在应用到缝合变形时保持恶意功能。 为了实际目的, 维护恶意功能是衡量对抗性攻击强度的重要尺度。 为了帮助 DNNS学会更彻底地防范攻击, 我们提议了一种亲异性对立性对称攻击攻击, 我们提议了一种通性对立性对立性攻击, 这种对立性攻击可以持续地产生更强的对立性对立实例。 为了效率, 我们提议将目前从Euclidean 几何测距的对立法协调平面的对平面图和对立面性模型进行大幅度的对准性变换, 也可以将原始的变异性实验方法, 将原始的变异性模型与任何变异性模型对等的对等方法, 。