We introduce Noise Injection Node Regularization (NINR), a method of injecting structured noise into Deep Neural Networks (DNN) during the training stage, resulting in an emergent regularizing effect. We present theoretical and empirical evidence for substantial improvement in robustness against various test data perturbations for feed-forward DNNs when trained under NINR. The novelty in our approach comes from the interplay of adaptive noise injection and initialization conditions such that noise is the dominant driver of dynamics at the start of training. As it simply requires the addition of external nodes without altering the existing network structure or optimization algorithms, this method can be easily incorporated into many standard problem specifications. We find improved stability against a number of data perturbations, including domain shifts, with the most dramatic improvement obtained for unstructured noise, where our technique outperforms other existing methods such as Dropout or $L_2$ regularization, in some cases. We further show that desirable generalization properties on clean data are generally maintained.
翻译:在培训阶段,我们引入了向深神经网络注入结构化噪音的一种方法,即向深神经网络注入结构化噪音,从而产生一种突发的正常化效果;我们提出了理论和经验证据,证明在根据NIN进行训练时,对向导调调制调制调制调制调制调制调制调制调制调制调制调制调制调制数据的各种测试数据扰动有了显著改善;我们的方法新颖之处来自适应性噪音注入和初始化条件的相互作用,例如噪音在培训开始时是动态的主要驱动因素;由于它只是要求在不改变现有网络结构或优化算法的情况下增加外部节点,因此这种方法很容易被纳入许多标准的问题规格中;我们发现,在一些数据扰动性干扰下,包括域变换,有了最显著的改进,使无结构化的噪音在有些情况下,我们的技术超越了其他现有方法,如辍学或$L_2美元的正规化。我们进一步表明,清洁数据一般保持理想的通用特性。