The classical development of neural networks has primarily focused on learning mappings between finite dimensional Euclidean spaces or finite sets. We propose a generalization of neural networks tailored to learn operators mapping between infinite dimensional function spaces. We formulate the approximation of operators by composition of a class of linear integral operators and nonlinear activation functions, so that the composed operator can approximate complex nonlinear operators. We prove a universal approximation theorem for our construction. Furthermore, we introduce four classes of operator parameterizations: graph-based operators, low-rank operators, multipole graph-based operators, and Fourier operators and describe efficient algorithms for computing with each one. The proposed neural operators are resolution-invariant: they share the same network parameters between different discretizations of the underlying function spaces and can be used for zero-shot super-resolutions. Numerically, the proposed models show superior performance compared to existing machine learning based methodologies on Burgers' equation, Darcy flow, and the Navier-Stokes equation, while being several order of magnitude faster compared to conventional PDE solvers.
翻译:典型的神经网络开发主要侧重于学习有限维度的Euclidean空间或有限数据集之间的测图。我们建议对神经网络进行一般化,专门为学习操作员绘制无限维功能空间之间的测图。我们通过一组线性整体操作员和非线性激活功能来对操作员进行近似,这样组成操作员就可以对复杂的非线性操作员进行近似。我们证明我们建造的“通用近似理论”。此外,我们引入了四类操作员参数化:基于图形的操作员、低级操作员、多极图性的操作员和Fourier操作员,并描述与每个操作员进行计算的有效算法。拟议的神经操作员是分辨率异性:它们共享基础功能空间不同离散的不同的网络参数,可用于零射超分辨率。从数字上看,拟议模型显示的性能优于基于布尔格斯方程式、达西流和纳维尔-斯托克斯方程式的现有机器学习方法,而与传统的PDE解算器相比,具有几等级级速度。