We develop a theoretical analysis for special neural network architectures, termed operator recurrent neural networks, for approximating nonlinear functions whose inputs are linear operators. Such functions commonly arise in solution algorithms for inverse boundary value problems. Traditional neural networks treat input data as vectors, and thus they do not effectively capture the multiplicative structure associated with the linear operators that correspond to the data in such inverse problems. We therefore introduce a new family that resembles a standard neural network architecture, but where the input data acts multiplicatively on vectors. Motivated by compact operators appearing in boundary control and the analysis of inverse boundary value problems for the wave equation, we promote structure and sparsity in selected weight matrices in the network. After describing this architecture, we study its representation properties as well as its approximation properties. We furthermore show that an explicit regularization can be introduced that can be derived from the mathematical analysis of the mentioned inverse problems, and which leads to certain guarantees on the generalization properties. We observe that the sparsity of the weight matrices improves the generalization estimates. Lastly, we discuss how operator recurrent networks can be viewed as a deep learning analogue to deterministic algorithms such as boundary control for reconstructing the unknown wavespeed in the acoustic wave equation from boundary measurements.
翻译:我们为特殊的神经网络结构(称为操作者经常性神经网络)进行理论分析,以接近投入为线性操作者的非线性功能。这些功能通常出现在反边界值问题的解决方案算法中。传统的神经网络将输入数据作为矢量处理,因此它们无法有效地捕捉线性操作者与此类反问题中的数据相对应的多复制结构。因此,我们引入了一个类似于标准神经网络结构的新组,但输入数据在矢量上具有倍增作用。在边界控制和分析波形反边界值问题时出现的紧凑操作者激励下,我们促进网络中选定重力矩阵的结构和宽度。在描述这一结构之后,我们研究其表达性及其近似性。我们进一步表明,可以从上述反问题数学分析中引入明确的规范化,从而导致对一般化特性的某些保证。我们观察到,重力矩阵的紧张性能改进了一般化估计。最后,我们讨论了操作者经常性网络在选定重力矩阵中的结构和宽度矩阵。在描述这一结构之后,我们研究其表达性特性特性的特性特性特性特性,从深度的边界测算法中将它视为从深度的测算为测算。