This paper studies the novel concept of weight correlation in deep neural networks and discusses its impact on the networks' generalisation ability. For fully-connected layers, the weight correlation is defined as the average cosine similarity between weight vectors of neurons, and for convolutional layers, the weight correlation is defined as the cosine similarity between filter matrices. Theoretically, we show that, weight correlation can, and should, be incorporated into the PAC Bayesian framework for the generalisation of neural networks, and the resulting generalisation bound is monotonic with respect to the weight correlation. We formulate a new complexity measure, which lifts the PAC Bayes measure with weight correlation, and experimentally confirm that it is able to rank the generalisation errors of a set of networks more precisely than existing measures. More importantly, we develop a new regulariser for training, and provide extensive experiments that show that the generalisation error can be greatly reduced with our novel approach.
翻译:本文研究了深神经网络中重力相关性的新概念,并讨论了其对网络总体化能力的影响。对于完全连接的层层,加权相关性被定义为神经质载体和进化层重量矢量之间的平均余弦相似性,而对于进化层,加权相关性被定义为过滤器基质之间的余弦相似性。从理论上讲,我们表明,重量相关性可以而且应该被纳入PAC Bayesian神经网络总体化框架,由此形成的一般化约束在重量相关性方面是单调的。我们制定了一个新的复杂度量度,将PAC Bayes测量值与重量相关性提升,并实验性地确认它能够比现有测量值更精确地排列一组网络的概括性错误。更重要的是,我们开发了一种新的常规化培训,并提供广泛的实验,表明通过我们的新方法,总化错误可以大大缩小。