Recent work has highlighted several advantages of enforcing orthogonality in the weight layers of deep networks, such as maintaining the stability of activations, preserving gradient norms, and enhancing adversarial robustness by enforcing low Lipschitz constants. Although numerous methods exist for enforcing the orthogonality of fully-connected layers, those for convolutional layers are more heuristic in nature, often focusing on penalty methods or limited classes of convolutions. In this work, we propose and evaluate an alternative approach to directly parameterize convolutional layers that are constrained to be orthogonal. Specifically, we propose to apply the Cayley transform to a skew-symmetric convolution in the Fourier domain, so that the inverse convolution needed by the Cayley transform can be computed efficiently. We compare our method to previous Lipschitz-constrained and orthogonal convolutional layers and show that it indeed preserves orthogonality to a high degree even for large convolutions. Applied to the problem of certified adversarial robustness, we show that networks incorporating the layer outperform existing deterministic methods for certified defense against $\ell_2$-norm-bounded adversaries, while scaling to larger architectures than previously investigated. Code is available at https://github.com/locuslab/orthogonal-convolutions.
翻译:最近的工作凸显了在深层网络重量层中强制执行正正数的几种优点,例如维持激活的稳定性,维护梯度规范,以及通过执行低利普西茨常量来增强对抗性强力。尽管有多种方法可以强制完全连接的层的正数性,但对于革命层来说,这些方法在性质上更为杂乱,往往侧重于惩罚方法或有限的演进类别。在这项工作中,我们提议并评价一种替代方法,直接参数化受约束或高调制约的卷流层。具体地说,我们提议将Cayley转换为四流域的Skew-对称性共振动,以便有效地计算Cayley变换所需的反向演动。我们比较我们的方法与以前的利普西茨受限制和高调的卷动层相比,并表明它确实保存着高度的正数性,甚至对于大型的卷流。适用于经认证的稳性强性问题。我们表明,网络在四流域域中将现有确定性平面的平流式共振动方法整合起来,同时对经认证的内压的正数级/直压结构结构进行比可测量。