Since the Lipschitz properties of CNN are widely considered to be related to adversarial robustness, we theoretically characterize the $\ell_1$ norm and $\ell_\infty$ norm of 2D multi-channel convolutional layers and provide efficient methods to compute the exact $\ell_1$ norm and $\ell_\infty$ norm. Based on our theorem, we propose a novel regularization method termed norm decay, which can effectively reduce the norms of convolutional layers and fully-connected layers. Experiments show that norm-regularization methods, including norm decay, weight decay, and singular value clipping, can improve generalization of CNNs. However, they can slightly hurt adversarial robustness. Observing this unexpected phenomenon, we compute the norms of layers in the CNNs trained with three different adversarial training frameworks and surprisingly find that adversarially robust CNNs have comparable or even larger layer norms than their non-adversarially robust counterparts. Furthermore, we prove that under a mild assumption, adversarially robust classifiers can be achieved, and can have an arbitrarily large Lipschitz constant. For this reason, enforcing small norms on CNN layers may be neither necessary nor effective in achieving adversarial robustness. The code is available at https://github.com/youweiliang/norm_robustness.
翻译:由于CNN的利普申茨特性被广泛认为与对抗性强力有关,因此,我们从理论上将2D多渠道共变层的2D多渠道共振层的规范与1美元标准与1美元标准与1美元标准相提并论,并提供计算准确的1美元标准与1美元标准的有效方法。根据我们的理论,我们建议一种新型的规范化方法,称为规范衰败,这可以有效减少共振层和完全相连层的规范化规范。实验表明,规范化方法,包括规范腐蚀、重量腐蚀和单值剪切,可以改善CNN的通用性。然而,它们可以略微伤害对抗性强强的规范。观察这种意想不到的现象,我们用三种不同的对抗性培训框架来计算CNN的规范。令人惊讶地发现,有敌意的CNNCNN的规范比非对抗性强强的对立层的规范可比甚至更大。此外,我们证明,在一种温和的假设下,可以实现稳健的分类,并且可以任意地使CNN得到大 Lipschitz/com的常态性。为此,因此,在CNN的规范中执行强势性规范。