In this paper, a new Discontinuity Capturing Shallow Neural Network (DCSNN) for approximating $d$-dimensional piecewise continuous functions and for solving elliptic interface problems is developed. There are three novel features in the present network; namely, (i) jump discontinuity is captured sharply, (ii) it is completely shallow consisting of only one hidden layer, (iii) it is completely mesh-free for solving partial differential equations (PDEs). We first continuously extend the $d$-dimensional piecewise continuous function in $(d+1)$-dimensional space by augmenting one coordinate variable to label the pieces of discontinuous function, and then construct a shallow neural network to express this new augmented function. Since only one hidden layer is employed, the number of training parameters (weights and biases) scales linearly with the dimension and the neurons used in the hidden layer. For solving elliptic interface equations, the network is trained by minimizing the mean squared error loss that consists of the residual of governing equation, boundary condition, and the interface jump conditions. We perform a series of numerical tests to compare the accuracy and efficiency of the present network. Our DCSNN model is comparably efficient due to only moderate number of parameters needed to be trained (a few hundreds of parameters used throughout all numerical examples here), and the result shows better accuracy (and less parameters) than other method using piecewise deep neural network in literature. We also compare the results obtained by the traditional grid-based immersed interface method (IIM) which is designed particularly for elliptic interface problems. Again, the present results show better accuracy than the ones obtained by IIM. We conclude by solving a six-dimensional problem to show the capability of the present network for high-dimensional applications.
翻译:在本文中, 开发了一个新的“ 不连续” 功能, 用于大约以美元为维面的 浅层神经网络( DCSNNN), 用于大约以美元+1美元为维面的直径连续函数, 并用于解决椭圆界面问题 。 在目前的网络中, 有三个新特点 。 即 (一) 跳跃不连续性被急剧捕获, (二) 它完全浅, 仅包含一个隐藏的层, (三) 解决部分差异方程式( PDE) 完全无网形。 我们首先通过增加一个坐标变量, 标定不连续功能的字符连续功能, 并随后构建一个浅线性神经网络, 来表达这个新的扩展功能。 由于只使用了一个隐藏的层, 培训参数( 重量和偏差) 由一个线性层组成, ( d+1) 以美元为维度的直径直径连续函数函数, 以$( d+ 1) 维空间为单位, 继续扩展一个坐标变量 的连续变量,, 并随后建一个浅度的神经网络 网络, 将当前 的精度 的精度 的精度 的精度测试显示 。