Previous work has cast doubt on the general framework of uniform convergence and its ability to explain generalization in neural networks. By considering a specific dataset, it was observed that a neural network completely misclassifies a projection of the training data (adversarial set), rendering any existing generalization bound based on uniform convergence vacuous. We provide an extensive theoretical investigation of the previously studied data setting through the lens of infinitely-wide models. We prove that the Neural Tangent Kernel (NTK) also suffers from the same phenomenon and we uncover its origin. We highlight the important role of the output bias and show theoretically as well as empirically how a sensible choice completely mitigates the problem. We identify sharp phase transitions in the accuracy on the adversarial set and study its dependency on the training sample size. As a result, we are able to characterize critical sample sizes beyond which the effect disappears. Moreover, we study decompositions of a neural network into a clean and noisy part by considering its canonical decomposition into its different eigenfunctions and show empirically that for too small bias the adversarial phenomenon still persists.
翻译:先前的工作使人们对统一的趋同总框架及其解释神经网络一般化的能力产生怀疑。 通过考虑具体的数据集,人们发现神经网络完全错误地对培训数据的预测(对立组合)进行分类,使任何现有的一般化在统一趋同的基础上是空洞的。我们通过无限范围模型的镜片对以前研究过的数据设置进行了广泛的理论调查。我们证明神经中枢(NTK)也存在同样的现象,我们发现其起源。我们强调产出偏差的重要作用,从理论上和从经验上表明明智的选择如何完全缓解问题。我们找出了对立组合准确性的尖锐阶段转变,并研究了其对培训样本大小的依赖性。结果,我们可以将影响消失的临界样本大小描述为更多。此外,我们研究神经网络进入清洁和噪音部分的分解,方法是考虑其神经分解成不同的二元功能,并从经验上表明对抗性现象仍然存在的偏差太小。