Deep Learning (DL), in particular deep neural networks (DNN), by default is purely data-driven and in general does not require physics. This is the strength of DL but also one of its key limitations when applied to science and engineering problems in which underlying physical properties and desired accuracy need to be achieved. DL methods in their original forms are not capable of respecting the underlying mathematical models or achieving desired accuracy even in big-data regimes. However, many data-driven science and engineering problems, such as inverse problems, typically have limited experimental or observational data, and DL would overfit the data in this case. Leveraging information encoded in the underlying mathematical models, we argue, not only compensates missing information in low data regimes but also provides opportunities to equip DL methods with the underlying physics, hence promoting better generalization. This paper develops a model-constrained deep learning approach and its variant TNet that are capable of learning information hidden in both the training data and the underlying mathematical models to solve inverse problems governed by partial differential equations. We provide the constructions and some theoretical results for the proposed approaches. We show that data randomization can enhance the smoothness of the networks and their generalizations. Comprehensive numerical results not only confirm the theoretical findings but also show that with even as little as 20 training data samples for 1D deconvolution, 50 for inverse 2D heat conductivity problem, 100 and 50 for inverse initial conditions for time-dependent 2D Burgers' equation and 2D Navier-Stokes equations, respectively. TNet solutions can be as accurate as Tikhonov solutions while being several orders of magnitude faster. This is possible owing to the model-constrained term, replications, and randomization.
翻译:深度学习( DL ), 特别是深层神经网络( DNN), 默认情况下, 默认情况下是纯数据驱动的, 一般不需要物理。 这是 DL 的力量, 但它也是在应用到科学和工程问题时的关键局限性之一, 而这些科学和工程问题需要实现基本的物理属性和期望的准确性。 DL 的原始形式方法无法尊重基本的数学模型或达到理想的准确性, 即使在大数据系统中也是如此。 然而, 许多数据驱动的科学和工程问题, 如反向问题, 通常只有有限的实验或观察数据, 而 DL 将在此情况下的数据配得过高。 在基本数学模型中编码的信息, 我们争辩说, D 不仅弥补了低数据系统中缺失的信息, 而且还提供了使 D 方法具备基本物理特性, 从而推动更精确化。 本文开发了一个模型集深层次的深层次学习方法及其变异TNet, 学习培训模型和基本数学模型中隐藏的信息, 解决部分差异方程方程式中的问题。 我们为最初的计算方法提供了构建和一些理论结果。, 。 数据在50 D 的模型化过程中, 显示, 将数据循环的模型化只能分别显示整个数据复制结果, 。