We present a novel loss formulation for efficient learning of complex dynamics from governing physics, typically described by partial differential equations (PDEs), using physics-informed neural networks (PINNs). In our experiments, existing versions of PINNs are seen to learn poorly in many problems, especially for complex geometries, as it becomes increasingly difficult to establish appropriate sampling strategy at the near boundary region. Overly dense sampling can adversely impede training convergence if the local gradient behaviors are too complex to be adequately modelled by PINNs. On the other hand, if the samples are too sparse, existing PINNs tend to overfit the near boundary region, leading to incorrect solution. To prevent such issues, we propose a new Boundary Connectivity (BCXN) loss function which provides linear local structure approximation (LSA) to the gradient behaviors at the boundary for PINN. Our BCXN-loss implicitly imposes local structure during training, thus facilitating fast physics-informed learning across entire problem domains with order of magnitude sparser training samples. This LSA-PINN method shows a few orders of magnitude smaller errors than existing methods in terms of the standard L2-norm metric, while using dramatically fewer training samples and iterations. Our proposed LSA-PINN does not pose any requirement on the differentiable property of the networks, and we demonstrate its benefits and ease of implementation on both multi-layer perceptron and convolutional neural network versions as commonly used in current PINN literature.
翻译:我们提出了一个新的损失配方,以便从物理治理中有效学习复杂的物理动态,典型的描述是部分差异方程(PDEs),使用物理知情神经网络(PINNs),我们提出一种新的损失配方。 在我们的实验中,现有版本的PINNs在许多问题中学习得不好,特别是复杂的地理分布,因为越来越难以在近边界区域制定适当的取样战略。过度密集的采样可能会妨碍培训的趋同,如果当地梯度行为过于复杂,无法由PINNs进行适当的模拟。另一方面,如果样本过于稀少,现有的PINNs往往过分适合近边界区域,导致错误的解决方案。在我们的实验中,我们建议新的边界连接(BXN)损失函数为PINNL边界上的梯度行为提供线性地方结构近似近(LSA)。我们的BXN损失在培训过程中隐含了当地结构,从而便利在整个问题领域迅速进行物理知情学习,其培训样本数量不小,而这种LSA-PINNNNN方法在标准版PR-NUR网络的执行方面,而我们使用的常规样本和可变式的内标定的模型和内定式的内建的内基内基内基内基内,则会大大地试验,其效益的好处减少。