Physics-Informed Neural Networks (PINN) are algorithms from deep learning leveraging physical laws by including partial differential equations together with a respective set of boundary and initial conditions as penalty terms into their loss function. In this work, we observe the significant role of correctly weighting the combination of multiple competitive loss functions for training PINNs effectively. To this end, we implement and evaluate different methods aiming at balancing the contributions of multiple terms of the PINNs loss function and their gradients. After reviewing of three existing loss scaling approaches (Learning Rate Annealing, GradNorm and SoftAdapt), we propose a novel self-adaptive loss balancing scheme for PINNs named \emph{ReLoBRaLo} (Relative Loss Balancing with Random Lookback). We extensively evaluate the performance of the aforementioned balancing schemes by solving both forward as well as inverse problems on three benchmark PDEs for PINNs: Burgers' equation, Kirchhoff's plate bending equation and Helmholtz's equation. The results show that ReLoBRaLo is able to consistently outperform the baseline of existing scaling methods in terms of accuracy, while also inducing significantly less computational overhead.
翻译:物理进化神经网络(PINN)是深层学习利用物理法法的算法,将部分差异方程式以及一套各自的边界和初始条件作为惩罚条件纳入损失功能中。 在这项工作中,我们观察到对多种竞争性损失功能的组合进行正确加权以有效培训PINNs的重要作用。为此,我们实施并评价了旨在平衡PINNs损失函数及其梯度的多重条件贡献的不同方法。在审查了三个现有的损失缩放方法(学习率安纳林、格拉德诺姆和SoftAdapt)之后,我们为名为\emph{Re Lobralo}的PINNs提出了一个新的自我调整损失平衡计划(与随机回放相调的相对应的补偿损失)。我们广泛评价了上述平衡计划的业绩,既向前又反向地解决了PINNs的三个基准PDEs:Burgers的方程式、Kirchhoff的板弯曲式方程式和Helmholtz的方程式。结果显示,ReLOBRALOLO在大幅提升现有基线条件的精确度的同时,也能够使现有基准条件的升级。