Xie Han, Xiong Li, Yang Carl
Emory University, Atlanta, GA, United States.
Proc ACM Int Conf Inf Knowl Manag. 2024;2024:2607-2617. doi: 10.1145/3627673.3679834. Epub 2024 Oct 21.
Federated learning on graphs (a.k.a., federated graph learning- FGL) has recently received increasing attention due to its capacity to enable collaborative learning over distributed graph datasets without compromising local clients' data privacy. In previous works, clients of FGL typically represent institutes or organizations that possess sets of entire graphs (e.g., molecule graphs in biochemical research) or parts of a larger graph (e.g., sub-user networks of e-commerce platforms). However, another natural paradigm exists where clients act as remote devices retaining the graph structures of local neighborhoods centered around the device owners (i.e., ego-networks), which can be modeled for specific graph applications such as user profiling on social ego-networks and infection prediction on contact ego-networks. FGL in such novel yet realistic ego-network settings faces the unique challenge of incomplete neighborhood information for non-ego local nodes since they likely appear and have different sets of neighbors in multiple ego-networks. To address this challenge, we propose an FGL method for distributed ego-networks in which clients obtain complete neighborhood information of local nodes through sharing node embeddings with other clients. A contrastive learning mechanism is proposed to bridge the gap between local and global node embeddings and stabilize the local training of graph neural network models, while a secure embedding sharing protocol is employed to protect individual node identity and embedding privacy against the server and other clients. Comprehensive experiments on various distributed ego-network datasets successfully demonstrate the effectiveness of our proposed embedding sharing method on top of different federated model sharing frameworks, and we also provide discussions on the potential efficiency and privacy drawbacks of the method as well as their future mitigation.
图上联邦学习(也称为联邦图学习 - FGL)最近受到越来越多的关注,因为它能够在不损害本地客户端数据隐私的情况下,对分布式图数据集进行协作学习。在先前的工作中,FGL的客户端通常代表拥有完整图集合(例如生化研究中的分子图)或更大图的一部分(例如电子商务平台的子用户网络)的机构或组织。然而,还存在另一种自然范式,即客户端充当远程设备,保留以设备所有者为中心的本地邻域的图结构(即自我网络),这可以针对特定的图应用进行建模,例如社交自我网络上的用户画像和接触自我网络上的感染预测。在这种新颖但现实的自我网络设置中的FGL面临着非自我本地节点邻域信息不完整的独特挑战,因为它们可能出现在多个自我网络中并且具有不同的邻居集。为了应对这一挑战,我们提出了一种用于分布式自我网络的FGL方法,其中客户端通过与其他客户端共享节点嵌入来获取本地节点的完整邻域信息。提出了一种对比学习机制来弥合本地和全局节点嵌入之间的差距,并稳定图神经网络模型的本地训练,同时采用一种安全的嵌入共享协议来保护单个节点身份和嵌入隐私,防止服务器和其他客户端获取。在各种分布式自我网络数据集上进行的综合实验成功证明了我们提出的嵌入共享方法在不同联邦模型共享框架之上的有效性,并且我们还讨论了该方法潜在的效率和隐私缺点以及它们未来的缓解措施。