Zhang Handi, Liu Langchen, Weng Kangyu, Lu Lu
IEEE Trans Neural Netw Learn Syst. 2025 Jun 27;PP. doi: 10.1109/TNNLS.2025.3580409.
By leveraging neural networks, the emerging field of scientific machine learning (SciML) offers novel approaches to address complex problems governed by partial differential equations (PDEs). In practical applications, challenges arise due to the distributed essence of data, concerns about data privacy, or the impracticality of transferring large volumes of data. Federated learning (FL), a decentralized framework that enables the collaborative training of a global model while preserving data privacy, offers a solution to the challenges posed by isolated data pools and sensitive data issues. Here, this article explores the integration of FL and SciML to approximate complex functions and solve differential equations. We propose two novel models: federated physics-informed neural networks (FedPINNs) and federated deep operator networks (FedDeepONets). We further introduce various data generation methods to control the degree of nonindependent and identically distributed (non-i.i.d.) data and utilize the 1-Wasserstein distance to quantify data heterogeneity in function approximation and PDE learning. We systematically investigate the relationship between data heterogeneity and federated model performance. In addition, we propose a measure of weight divergence and develop a theoretical framework to establish growth bounds for weight divergence in FL compared with centralized learning. To demonstrate the effectiveness of our methods, we conducted ten experiments, including two on function approximation, five PDE problems on FedPINN, and four PDE problems on FedDeepONet. These experiments demonstrate that proposed federated methods surpass the models trained only using local data and achieve competitive accuracy of centralized models trained using all data.
通过利用神经网络,新兴的科学机器学习(SciML)领域提供了新颖的方法来解决由偏微分方程(PDE)控制的复杂问题。在实际应用中,由于数据的分布式本质、对数据隐私的担忧或传输大量数据的不切实际性,会出现各种挑战。联邦学习(FL)是一种去中心化框架,能够在保护数据隐私的同时对全局模型进行协同训练,为孤立数据池和敏感数据问题带来的挑战提供了解决方案。在此,本文探讨了FL与SciML的集成,以逼近复杂函数并求解微分方程。我们提出了两种新颖的模型:联邦物理信息神经网络(FedPINNs)和联邦深度算子网络(FedDeepONets)。我们进一步引入了各种数据生成方法来控制非独立同分布(non-i.i.d.)数据的程度,并利用1-瓦瑟斯坦距离来量化函数逼近和PDE学习中的数据异质性。我们系统地研究了数据异质性与联邦模型性能之间的关系。此外,我们提出了一种权重散度度量,并开发了一个理论框架,以建立与集中式学习相比FL中权重散度的增长界限。为了证明我们方法的有效性,我们进行了十次实验,包括两次函数逼近实验、五次关于FedPINN的PDE问题实验和四次关于FedDeepONet的PDE问题实验。这些实验表明,所提出的联邦方法超越了仅使用本地数据训练的模型,并实现了使用所有数据训练的集中式模型的竞争精度。