Current methods for regularization in machine learning require quite specific model assumptions (e.g. a kernel shape) that are not derived from prior knowledge about the application, but must be imposed merely to make the method work. We show in this paper that regularization can indeed be achieved by assuming nothing but invariance principles (w.r.t. scaling, translation, and rotation of input and output space) and the degree of differentiability of the true function. Concretely, we derive a novel (non-Gaussian) stochastic process from the above minimal assumptions, and we present a corresponding Bayesian inference method for regression. The mean posterior turns out to be a polyharmonic spline, and the posterior process is a mixture of t-processes. Compared with Gaussian process regression, the proposed method shows equal performance and has the advantages of being (i) less arbitrary (no choice of kernel) (ii) potentially faster (no kernel parameter optimization), and (iii) having better extrapolation behavior. We believe that the proposed theory has central importance for the conceptual foundations of regularization and machine learning and also has great potential to enable practical advances in ML areas beyond regression.
翻译:目前机器学习的正规化方法要求非常具体的模型假设(如内核形状),这些假设并非来自对应用的先前知识,而必须强制施用,只是为了使方法发挥作用。我们在本文件中表明,正规化确实可以通过假设不折不扣的原则(W.r.t. 缩放、翻译和输入和输出空间的轮换)和真实功能的不同程度来实现。具体地说,我们从上述最低假设中得出了一个新颖的(非Gausian)随机过程,我们提出了相应的巴耶斯推论回归方法。平均的后表象结果显示为多合力螺纹,而后表象过程则是t-过程的混合。与高斯进程回归相比,拟议方法表现出同等的性能,其优点是(一)不那么武断(不选择内核);(二)可能更快(诺内尔参数优化),以及(三)具有更好的外推行为。我们认为,拟议的理论对于常规化和机器学习领域的概念基础而言具有核心重要性,而且使M在实际回归领域具有巨大的潜力。