Adversarial attacks make their success in DNNs, and among them, gradient-based algorithms become one of the mainstreams. Based on the linearity hypothesis, under $\ell_\infty$ constraint, $sign$ operation applied to the gradients is a good choice for generating perturbations. However, side-effects from such operation exist since it leads to the bias of direction between real gradients and perturbations. In other words, current methods contain a gap between real gradients and actual noises, which leads to biased and inefficient attacks. Therefore in this paper, based on the Taylor expansion, the bias is analyzed theoretically, and the correction of $sign$, i.e., Fast Gradient Non-sign Method (FGNM), is further proposed. Notably, FGNM is a general routine that seamlessly replaces the conventional $sign$ operation in gradient-based attacks with negligible extra computational cost. Extensive experiments demonstrate the effectiveness of our methods. Specifically, for untargeted black-box attacks, ours outperform them by 27.5% at most and 9.5% on average. For targeted attacks against defense models, it is 15.1% and 12.7%. Our anonymous code is publicly available at https://github.com/yaya-cheng/FGNM
翻译:自动攻击在DNNS中取得成功,其中,梯度式算法成为主流之一。根据线性假设,在$\ell\ ⁇ infty美元的限制下,对梯度应用的美元操作是产生扰动的好选择。然而,这种操作的副作用是存在的,因为它导致真实梯度和扰动之间方向偏差。换句话说,目前的方法包含实际梯度与实际噪音之间的差距,导致偏差和低效攻击。因此,在本文中,根据泰勒扩张,从理论上分析了偏差,并进一步提出了美元(即快速梯度非特配法)的更正。值得注意的是,FGNM是一种普通的例行做法,它以微不足道的计算成本无缝取代了基于梯度的攻击中常规的美元操作。广泛的实验显示了我们的方法的有效性。具体地说,对于目标不明确的黑箱攻击,我们的方法比它们高出了27.5%,平均为9.5%。对于目标性攻击,即快速梯度非特制方法(即快速非特制方法)方法(FNMM)是15.1%/正式的公开防御模式。我们使用的是15.1%/正版的M.