We develop a variational framework to understand the properties of the functions learned by neural networks fit to data. We propose and study a family of continuous-domain linear inverse problems with total variation-like regularization in the Radon domain subject to data fitting constraints. We derive a representer theorem showing that finite-width, single-hidden layer neural networks are solutions to these inverse problems. We draw on many techniques from variational spline theory and so we propose the notion of polynomial ridge splines, which correspond to single-hidden layer neural networks with truncated power functions as the activation function. The representer theorem is reminiscent of the classical reproducing kernel Hilbert space representer theorem, but we show that the neural network problem is posed over a non-Hilbertian Banach space. While the learning problems are posed in the continuous-domain, similar to kernel methods, the problems can be recast as finite-dimensional neural network training problems. These neural network training problems have regularizers which are related to the well-known weight decay and path-norm regularizers. Thus, our result gives insight into functional characteristics of trained neural networks and also into the design neural network regularizers. We also show that these regularizers promote neural network solutions with desirable generalization properties.
翻译:我们开发了一个可变框架, 以了解神经网络所学功能的特性。 我们提议并研究一个连续的多面线性反向问题, 在受数据安装制约的情况下, 在 Radon 域中存在完全变异的正规化问题。 我们产生一个代表理论, 显示有限宽度、 单层层神经网络是这些反向问题的解决方案。 我们从变异样样样样理论中吸取了许多技术, 因此我们建议了多面脊样条的理念, 与单层层神经网络相对应, 其电源功能为激活功能。 代表的理论是古典复制核心内尔· 希尔伯特 空域代表理论的类似。 但我们显示, 神经网络问题是由非 Hilbertian Banach 空间造成的。 虽然学习问题在连续多面中存在, 类似内核方法, 问题可以被重新表述为定式神经网络培训问题。 这些神经网络培训问题与已知的神经网络的正统化性解决方案有关, 也让我们的正统化性网络路径和系统显示这些常态的系统。