Scientific Computing and Imaging Institute, University of Utah, Salt Lake City, UT 84112, USA; Kahlert School of Computing, University of Utah, Salt Lake City, UT 84112, USA.
Department of Computing and Mathematical Sciences, Caltech, Pasadena, CA 91125, USA.
Neural Netw. 2024 Dec;180:106703. doi: 10.1016/j.neunet.2024.106703. Epub 2024 Sep 4.
Physics-informed machine learning (PIML) as a means of solving partial differential equations (PDEs) has garnered much attention in the Computational Science and Engineering (CS&E) world. This topic encompasses a broad array of methods and models aimed at solving a single or a collection of PDE problems, called multitask learning. PIML is characterized by the incorporation of physical laws into the training process of machine learning models in lieu of large data when solving PDE problems. Despite the overall success of this collection of methods, it remains incredibly difficult to analyze, benchmark, and generally compare one approach to another. Using Kolmogorov n-widths as a measure of effectiveness of approximating functions, we judiciously apply this metric in the comparison of various multitask PIML architectures. We compute lower accuracy bounds and analyze the model's learned basis functions on various PDE problems. This is the first objective metric for comparing multitask PIML architectures and helps remove uncertainty in model validation from selective sampling and overfitting. We also identify avenues of improvement for model architectures, such as the choice of activation function, which can drastically affect model generalization to "worst-case" scenarios, which is not observed when reporting task-specific errors. We also incorporate this metric into the optimization process through regularization, which improves the models' generalizability over the multitask PDE problem.
物理信息机器学习(PIML)作为求解偏微分方程(PDE)的一种手段,在计算科学与工程(CS&E)领域引起了广泛关注。这个主题涵盖了广泛的方法和模型,旨在解决一个或一组 PDE 问题,称为多任务学习。PIML 的特点是在解决 PDE 问题时,将物理定律纳入机器学习模型的训练过程中,而不是依赖大量数据。尽管这些方法总体上取得了成功,但分析、基准测试和一般比较一种方法与另一种方法仍然非常困难。我们使用 Kolmogorov n 宽度作为函数逼近效果的度量标准,在比较各种多任务 PIML 架构时明智地应用了这一指标。我们计算了较低的精度边界,并分析了模型在各种 PDE 问题上学习的基函数。这是比较多任务 PIML 架构的第一个客观指标,并有助于消除模型验证中因选择性采样和过拟合而导致的不确定性。我们还确定了模型架构的改进途径,例如激活函数的选择,这会极大地影响模型对“最坏情况”场景的泛化能力,而在报告特定任务的误差时则不会观察到这种情况。我们还通过正则化将这个指标纳入到优化过程中,从而提高了模型在多任务 PDE 问题上的泛化能力。