Kumar Varun, Goswami Somdatta, Kontolati Katiana, Shields Michael D, Karniadakis George Em
School of Engineering, Brown University, United States of America.
Department of Civil and Systems Engineering, Johns Hopkins University, United States of America.
Neural Netw. 2025 Apr;184:107113. doi: 10.1016/j.neunet.2024.107113. Epub 2025 Jan 3.
Multi-task learning (MTL) is an inductive transfer mechanism designed to leverage useful information from multiple tasks to improve generalization performance compared to single-task learning. It has been extensively explored in traditional machine learning to address issues such as data sparsity and overfitting in neural networks. In this work, we apply MTL to problems in science and engineering governed by partial differential equations (PDEs). However, implementing MTL in this context is complex, as it requires task-specific modifications to accommodate various scenarios representing different physical processes. To this end, we present a multi-task deep operator network (MT-DeepONet) to learn solutions across various functional forms of source terms in a PDE and multiple geometries in a single concurrent training session. We introduce modifications in the branch network of the vanilla DeepONet to account for various functional forms of a parameterized coefficient in a PDE. Additionally, we handle parameterized geometries by introducing a binary mask in the branch network and incorporating it into the loss term to improve convergence and generalization to new geometry tasks. Our approach is demonstrated on three benchmark problems: (1) learning different functional forms of the source term in the Fisher equation; (2) learning multiple geometries in a 2D Darcy Flow problem and showcasing better transfer learning capabilities to new geometries; and (3) learning 3D parameterized geometries for a heat transfer problem and demonstrate the ability to predict on new but similar geometries. Our MT-DeepONet framework offers a novel approach to solving PDE problems in engineering and science under a unified umbrella based on synergistic learning that reduces the overall training cost for neural operators.
多任务学习(MTL)是一种归纳迁移机制,旨在利用来自多个任务的有用信息,以提高与单任务学习相比的泛化性能。它在传统机器学习中已被广泛探索,以解决诸如神经网络中的数据稀疏性和过拟合等问题。在这项工作中,我们将多任务学习应用于由偏微分方程(PDE)支配的科学和工程问题。然而,在这种情况下实现多任务学习很复杂,因为它需要针对特定任务进行修改,以适应代表不同物理过程的各种场景。为此,我们提出了一种多任务深度算子网络(MT-DeepONet),以便在单个并发训练会话中学习偏微分方程中各种源项函数形式和多种几何形状的解。我们对普通深度算子网络的分支网络进行了修改,以考虑偏微分方程中参数化系数的各种函数形式。此外,我们通过在分支网络中引入二进制掩码并将其纳入损失项来处理参数化几何形状,以提高收敛性并推广到新的几何形状任务。我们的方法在三个基准问题上得到了验证:(1)学习费希尔方程中源项的不同函数形式;(2)学习二维达西流问题中的多种几何形状,并展示对新几何形状更好的迁移学习能力;(3)学习三维参数化几何形状的热传导问题,并展示对新的但相似几何形状进行预测的能力。我们的MT-DeepONet框架提供了一种基于协同学习的统一方法来解决工程和科学中的偏微分方程问题,从而降低了神经算子的总体训练成本。