Lipschitz constrained models have been used to solve specifics deep learning problems such as the estimation of Wasserstein distance for GAN, or the training of neural networks robust to adversarial attacks. Regardless the novel and effective algorithms to build such 1-Lipschitz networks, their usage remains marginal, and they are commonly considered as less expressive and less able to fit properly the data than their unconstrained counterpart. The goal of the paper is to demonstrate that, despite being empirically harder to train, 1-Lipschitz neural networks are theoretically better grounded than unconstrained ones when it comes to classification. To achieve that we recall some results about 1-Lipschitz function in the scope of deep learning and we extend and illustrate them to derive general properties for classification. First, we show that 1-Lipschitz neural network can fit arbitrarily difficult frontier making them as expressive as classical ones. When minimizing the log loss, we prove that the optimization problem under Lipschitz constraint is well posed and have a minimum, whereas regular neural networks can diverge even on remarkably simple situations. Then, we study the link between classification with 1-Lipschitz network and optimal transport thanks to regularized versions of Kantorovich-Rubinstein duality theory. Last, we derive preliminary bounds on their VC dimension.
翻译:Lipschitz 限制模型被用来解决深层学习的具体问题,例如,估计瓦塞斯坦距离GAN的距离,或培训对对抗性攻击具有强大的神经网络。不管建立这种1-利普西茨网络的新颖而有效的算法,它们的使用仍然微不足道,而且通常认为这些算法比没有受到约束的对等网络更不那么清晰,而且不那么适合数据。本文的目的是要表明,尽管在经验上难以培训,但1-利普西茨神经网络在理论上比在分类时不受限制的网络在基础上要好一些。为了实现这一点,我们在深层次学习范围内回顾了关于1-利普西茨功能的一些结果,我们扩展和说明这些结果以获得一般属性。首先,我们表明,1-利普西茨神经网络可以任意地适合困难的边界,使其像古典的对口一样表达。当最大限度地减少日志损失时,我们证明利普西茨约束下的优化问题已经很好地形成,并且有一个最起码的,而常规的神经网络在非常简单的情况下也可能是不同的。然后,我们研究与1-利普西茨初步理论化的系统网络和我们最优化的卡什维新版本的系统之间的连接。