Deep neural networks can be fragile and sensitive to small input perturbations that might cause a significant change in the output. In this paper, we employ contraction theory to improve the robustness of neural ODEs (NODEs). A dynamical system is contractive if all solutions with different initial conditions converge to each other exponentially fast. As a consequence, perturbations in initial conditions become less and less relevant over time. Since in NODEs the input data corresponds to the initial condition of dynamical systems, we show contractivity can mitigate the effect of input perturbations. More precisely, inspired by NODEs with Hamiltonian dynamics, we propose a class of contractive Hamiltonian NODEs (CH-NODEs). By properly tuning a scalar parameter, CH-NODEs ensure contractivity by design and can be trained using standard backpropagation. Moreover, CH-NODEs enjoy built-in guarantees of non-exploding gradients, which ensure a well-posed training process. Finally, we demonstrate the robustness of CH-NODEs on the MNIST image classification problem with noisy test data.
翻译:深神经网络可能脆弱,对小输入扰动敏感,这可能导致产出的显著变化。在本文中,我们采用收缩理论来提高神经值的稳健性。如果最初条件不同的所有解决方案相互交汇,动态系统就会以指数速度加速。因此,初始条件下的扰动随着时间推移变得越来越少,也变得不那么相关。由于输入数据与动态系统初始状态相对应,我们显示,收缩性可以减轻输入振动的效果。更准确地说,在汉密尔顿动态的NODs的启发下,我们提出了一组压缩的汉密尔顿耐克模式。通过适当调整一个标度参数,CH-NODs通过设计确保合同性,并且可以使用标准的反演算方法进行培训。此外,CH-NODEs享有非爆炸性梯度的内在保证,这确保了良好的培训过程。最后,我们展示了CH-NOD在MNIST图像分类上与噪音测试数据有关的问题的稳健性。