Data sparsity is a common issue to train machine learning tools such as neural networks for engineering and scientific applications, where experiments and simulations are expensive. Recently physics-constrained neural networks (PCNNs) were developed to reduce the required amount of training data. However, the weights of different losses from data and physical constraints are adjusted empirically in PCNNs. In this paper, a new physics-constrained neural network with the minimax architecture (PCNN-MM) is proposed so that the weights of different losses can be adjusted systematically. The training of the PCNN-MM is searching the high-order saddle points of the objective function. A novel saddle point search algorithm called Dual-Dimer method is developed. It is demonstrated that the Dual-Dimer method is computationally more efficient than the gradient descent ascent method for nonconvex-nonconcave functions and provides additional eigenvalue information to verify search results. A heat transfer example also shows that the convergence of PCNN-MMs is faster than that of traditional PCNNs.
翻译:数据宽度是培训机械学习工具,如神经网络等用于工程和科学应用、实验和模拟费用昂贵的神经网络的一个常见问题。最近,为减少所需培训数据量,开发了物理学限制的神经网络(PCNNN),但数据和物理限制造成的不同损失的权重在PCNNN中进行了实验性调整。在本文中,提议建立一个新的物理学限制的神经网络,使用微型数学结构(PCNNN-MM),以便系统地调整不同损失的权重。PCNN-MM的培训正在搜索目标功能的高阶轮廓。开发了一种称为双二模方法的新型马鞍搜索算法。这表明,双维调方法在计算上比非convex-nonconcave功能的梯度下降法效率更高,并且提供了额外的电子价值信息以核实搜索结果。一个热传输的例子还表明,PCNNN-MM的趋同率比传统的PCNNNN的速度要快。