Several recent works in scientific machine learning have revived interest in the application of neural networks to partial differential equations (PDEs). A popular approach is to aggregate the residual form of the governing PDE and its boundary conditions as soft penalties into a composite objective/loss function for training neural networks, which is commonly referred to as physics-informed neural networks (PINNs). In the present study, we visualize the loss landscapes and distributions of learned parameters and explain the ways this particular formulation of the objective function may hinder or even prevent convergence when dealing with challenging target solutions. We construct a purely data-driven loss function composed of both the boundary loss and the domain loss. Using this data-driven loss function and, separately, a physics-informed loss function, we then train two neural network models with the same architecture. We show that incomparable scales between boundary and domain loss terms are the culprit behind the poor performance. Additionally, we assess the performance of both approaches on two elliptic problems with increasingly complex target solutions. Based on our analysis of their loss landscapes and learned parameter distributions, we observe that a physics-informed neural network with a composite objective function formulation produces highly non-convex loss surfaces that are difficult to optimize and are more prone to the problem of vanishing gradients.
翻译:科学机器学习方面的最近几项工程重新唤起人们对将神经网络应用于局部差异方程式的兴趣。一种流行的做法是将管辖的PDE及其边界条件的剩余形式作为软惩罚,汇总成培训神经网络的综合目标/损失功能,通常称为物理知情神经网络。在本研究中,我们设想了损失的景象和所学参数的分布,并解释了在应对具有挑战性的目标解决方案时,这一目标功能的特定表述可能阻碍甚至防止趋同的方法。我们建立了一个纯数据驱动的损失功能,由边界损失和域损失组成。我们利用这一数据驱动的损失功能,并单独利用一个物理知情的损失功能,我们然后用同一结构来培训两个神经网络模型。我们表明,边界和域损失术语之间在可比较的尺度上是表现不佳的罪魁祸首。此外,我们评估了两种具有日益复杂目标解决方案的椭圆形问题两种方法的性能。我们根据对损失方貌和所学参数分布的分析,我们发现,一个基于物理定位的神经网络,具有最难的、最易降解的地平面功能,造成高度的不易变的地层。