Farhani Ghazal, Dashtbayaz Nima Hosseini, Kazachek Alexander, Wang Boyu
National Research Council Canada, Automotive and Surface Transportation, 800 Collip Cir, London, N6G 4X8, Canada.
Western University, Department of Computer Science, 1151 Richmond St, Middlesex College, London, N6A 5B7, Canada.
Neural Netw. 2025 Mar;183:106963. doi: 10.1016/j.neunet.2024.106963. Epub 2024 Dec 10.
Physics-informed neural networks (PINNs) have shown promising results in solving a wide range of problems involving partial differential equations (PDEs). Nevertheless, there are several instances of the failure of PINNs when PDEs become more complex. Particularly, when PDE coefficients grow larger or PDEs become increasingly nonlinear, PINNs struggle to converge to the true solution. A noticeable discrepancy emerges in the convergence speed between the PDE loss and the initial/boundary conditions loss, leading to the inability of PINNs to effectively learn the true solutions to these PDEs. In the present work, leveraging the neural tangent kernels (NTKs), we investigate the training dynamics of PINNs. Our theoretical analysis reveals that when PINNs are trained using gradient descent with momentum (GDM), the gap in convergence rates between the two loss terms is significantly reduced, thereby enabling the learning of the exact solution. We also examine why training a model via the Adam optimizer can accelerate the convergence and reduce the effect of the mentioned discrepancy. Our numerical experiments validate that sufficiently wide networks trained with GDM and Adam yield desirable solutions for more complex PDEs.
物理信息神经网络(PINNs)在解决涉及偏微分方程(PDEs)的广泛问题方面已显示出有前景的结果。然而,当PDEs变得更加复杂时,PINNs会有若干失败的情况。特别是,当PDE系数变得更大或PDEs变得越来越非线性时,PINNs难以收敛到真实解。在PDE损失和初始/边界条件损失之间的收敛速度出现明显差异,导致PINNs无法有效地学习这些PDEs的真实解。在当前工作中,利用神经切线核(NTKs),我们研究了PINNs的训练动态。我们的理论分析表明,当使用带动量的梯度下降(GDM)训练PINNs时,两个损失项之间的收敛速率差距会显著减小,从而能够学习到精确解。我们还研究了为什么通过Adam优化器训练模型可以加速收敛并减少上述差异的影响。我们的数值实验验证了用GDM和Adam训练的足够宽的网络能为更复杂的PDEs产生理想的解。