We develop a variational framework to understand the properties of the functions learned by neural networks fit to data. We propose and study a family of continuous-domain linear inverse problems with total variation-like regularization in the Radon domain subject to data fitting constraints. We derive a representer theorem showing that finite-width, single-hidden layer neural networks are solutions to these inverse problems. We draw on many techniques from variational spline theory and so we propose the notion of polynomial ridge splines, which correspond to a single-hidden layer neural networks with truncated power functions as the activation function. The representer theorem is reminiscent of the classical reproducing kernel Hilbert space representer theorem, but we show that the neural network problem is posed over a non-Hilbertian Banach space. While the learning problems are posed in the continuous-domain, similar to kernel methods, the problems can be recast as finite-dimensional neural network training problems. These neural network training problems have regularizers which are related to the well-known weight decay and path-norm regularizers. Thus, our result gives insight into functional characteristics of trained neural networks and also into the design neural network regularizers. We also show that these regularizers promote neural network solutions with desirable generalization properties.
翻译:我们开发了一个可变框架, 以了解神经网络所学功能的特性。 我们提议并研究一个连续的多面线性反向问题, 以在适应数据的限制下, 在 Radon 域中出现完全变异的正规化问题。 我们产生一个代表理论, 显示有限宽度、 单层层神经网络是这些反向问题的解决方案。 我们从变式样样理论中吸取了许多技术, 因此我们建议了多面脊样条的理念, 与一个单层层神经网络相对应, 其电源功能是流动的, 作为激活功能。 代表的理论是古典复制内核Hilbert 空间代表理论的象征, 但我们显示神经网络问题是由非Hilbert Banach 空间造成的。 虽然学习问题是在连续的多面上产生的, 类似于内核方法, 问题可以被重新表述为一维度神经网络的培训问题。 这些神经网络培训问题具有调和常规神经网络特性, 也使得常规的神经网络的系统化过程。