We propose a theoretical understanding of neural networks in terms of Wilsonian effective field theory. The correspondence relies on the fact that many asymptotic neural networks are drawn from Gaussian processes, the analog of non-interacting field theories. Moving away from the asymptotic limit yields a non-Gaussian process and corresponds to turning on particle interactions, allowing for the computation of correlation functions of neural network outputs with Feynman diagrams. Minimal non-Gaussian process likelihoods are determined by the most relevant non-Gaussian terms, according to the flow in their coefficients induced by the Wilsonian renormalization group. This yields a direct connection between overparameterization and simplicity of neural network likelihoods. Whether the coefficients are constants or functions may be understood in terms of GP limit symmetries, as expected from 't Hooft's technical naturalness. General theoretical calculations are matched to neural network experiments in the simplest class of models allowing the correspondence. Our formalism is valid for any of the many architectures that becomes a GP in an asymptotic limit, a property preserved under certain types of training.
翻译:我们建议从理论角度理解神经网络, 即威尔逊有效的实地理论。 函文依据这一事实, 许多微量神经网络来自高森过程, 类似于非互动的实地理论。 离开无症状限制可以产生一种非高加索过程, 并相对应于转向粒子相互作用, 允许用费曼图计算神经网络输出的关联功能。 最小的非加西语过程可能性是由最相关的非加西语术语决定的。 根据威尔逊再正常化组引申的系数流, 我们的正规主义可以适用于在神经网络可能性的简单化中成为GP的众多结构。 参数是常数还是函数, 可以按照“ 霍夫特” 技术自然性所预期的GP限制对称来理解。 一般理论计算与允许通信的最简单模型中的神经网络实验相匹配。 我们的正规主义对于在某种程度的训练中成为GP的架构, 在某种程度的限制下保存属性类型。