Randomly perturbing networks during the training process is a commonly used approach to improving generalization performance. In this paper, we present a theoretical study of one particular way of random perturbation, which corresponds to injecting artificial noise to the training data. We provide a precise asymptotic characterization of the training and generalization errors of such randomly perturbed learning problems on a random feature model. Our analysis shows that Gaussian noise injection in the training process is equivalent to introducing a weighted ridge regularization, when the number of noise injections tends to infinity. The explicit form of the regularization is also given. Numerical results corroborate our asymptotic predictions, showing that they are accurate even in moderate problem dimensions. Our theoretical predictions are based on a new correlated Gaussian equivalence conjecture that generalizes recent results in the study of random feature models.
翻译:培训过程中的随机扰动网络是一种常用的方法,用来改进一般化绩效。本文介绍对随机扰动的一种特定方式的理论研究,这与培训数据中注入人工噪音相对应。我们在随机特征模型中对随机扰动学习问题的培训和一般化错误作了精确的简单描述。我们的分析表明,在培训过程中,高山噪音注入相当于引入加权脊固定化,因为噪音注入的数量往往具有无限性。还给出了规范化的明确形式。数字结果证实了我们的无症状预测,表明即使在中度问题方面,这些预测也是准确的。我们的理论预测基于一个新的相互关联的高山等值等值预测,该预测概括了随机特征模型研究的最新结果。