Assumptions about invariances or symmetries in data can significantly increase the predictive power of statistical models. Many commonly used models in machine learning are constraint to respect certain symmetries in the data, such as translation equivariance in convolutional neural networks, and incorporation of new symmetry types is actively being studied. Yet, efforts to learn such invariances from the data itself remains an open research problem. It has been shown that marginal likelihood offers a principled way to learn invariances in Gaussian Processes. We propose a weight-space equivalent to this approach, by minimizing a lower bound on the marginal likelihood to learn invariances in neural networks resulting in naturally higher performing models.
翻译:关于数据差异或对称的假设可以大大增加统计模型的预测力。许多在机器学习中常用的模式都制约着尊重数据的某些对称性,如在神经神经网络中翻译的对称性,正在积极研究新的对称类型。然而,从数据本身中了解这种差异的努力仍然是一个开放的研究问题。已经证明,边际可能性提供了一种原则性的方法来学习高斯进程中的不差异性。我们建议了一个与这一方法相当的权重空间,最大限度地降低在神经网络中学习导致自然更高性能模型的不一致性的边际范围。