Recently, Zhang et al. (2021) developed a new neural network architecture based on $\ell_\infty$-distance functions, which naturally possesses certified robustness by its construction. Despite the excellent theoretical properties, the model so far can only achieve comparable performance to conventional networks. In this paper, we significantly boost the certified robustness of $\ell_\infty$-distance nets through a careful analysis of its training process. In particular, we show the $\ell_p$-relaxation, a crucial way to overcome the non-smoothness of the model, leads to an unexpected large Lipschitz constant at the early training stage. This makes the optimization insufficient using hinge loss and produces sub-optimal solutions. Given these findings, we propose a simple approach to address the issues above by using a novel objective function that combines a scaled cross-entropy loss with clipped hinge loss. Our experiments show that using the proposed training strategy, the certified accuracy of $\ell_\infty$-distance net can be dramatically improved from 33.30% to 40.06% on CIFAR-10 ($\epsilon=8/255$), meanwhile significantly outperforming other approaches in this area. Such a result clearly demonstrates the effectiveness and potential of $\ell_\infty$-distance net for certified robustness.
翻译:最近,张等人(2021年)开发了一个新的神经网络结构,其基础是$ell ⁇ infty-lear 函数,这种功能自然地具有经认证的稳健性。尽管理论特性极好,但迄今为止该模型只能达到与常规网络的类似性能。在本文中,我们通过仔细分析其培训过程,大大提升了经认证的美元/infty-le网的稳健性。特别是,我们展示了美元/p$-relax,这是克服该模型不顺差的关键方法,导致在早期培训阶段出现出乎意料的大型利普施奇茨常态。这使得利用断链损失和产生亚最佳解决方案的优化不足。根据这些发现,我们建议了一种简单的方法来解决上述问题,即使用一种新的目标功能,将规模的跨翼损失与剪接的链条损失结合起来。我们的实验表明,使用拟议的培训战略,经认证的美元/inty-lenet净额的准确性能大大改进,从33.0 %到40.6%/美元/eepslon==255美元,从而明显展示了这个区域的潜在效益。