Mai Chengyuan, Chang Yaomin, Chen Chuan, Zheng Zibin
IEEE Trans Neural Netw Learn Syst. 2025 Jan;36(1):1258-1271. doi: 10.1109/TNNLS.2023.3333846. Epub 2025 Jan 7.
Graph neural networks (GNNs) have achieved state-of-the-art performance in various graph representation learning scenarios. However, when applied to graph data in real world, GNNs have encountered scalability issues. Existing GNNs often have high computational load in both training and inference stages, making them incapable of meeting the performance needs of large-scale scenarios with a large number of nodes. Although several studies on scalable GNNs have developed, they either merely improve GNNs with limited scalability or come at the expense of reduced effectiveness. Inspired by knowledge distillation's (KDs) achievement in preserving performances while balancing scalability in computer vision and natural language processing, we propose an enhanced scalable GNN via KD (KD-SGNN) to improve the scalability and effectiveness of GNNs. On the one hand, KD-SGNN adopts the idea of decoupled GNNs, which decouples feature transformation and feature propagation in GNNs and leverages preprocessing techniques to improve the scalability of GNNs. On the other hand, KD-SGNN proposes two KD mechanisms (i.e., soft-target (ST) distillation and shallow imitation (SI) distillation) to improve the expressiveness. The scalability and effectiveness of KD-SGNN are evaluated on multiple real datasets. Besides, the effectiveness of the proposed KD mechanisms is also verified through comprehensive analyses.
图神经网络(GNNs)在各种图表示学习场景中取得了领先的性能。然而,当应用于现实世界的图数据时,GNNs遇到了可扩展性问题。现有的GNNs在训练和推理阶段通常具有很高的计算负载,使其无法满足具有大量节点的大规模场景的性能需求。尽管已经开展了几项关于可扩展GNNs的研究,但它们要么只是以有限的可扩展性改进GNNs,要么是以降低有效性为代价。受知识蒸馏(KD)在计算机视觉和自然语言处理中平衡可扩展性的同时保持性能方面的成就启发,我们提出了一种通过KD的增强型可扩展GNN(KD-SGNN),以提高GNNs的可扩展性和有效性。一方面,KD-SGNN采用了解耦GNNs的思想,它将GNNs中的特征变换和特征传播解耦,并利用预处理技术来提高GNNs的可扩展性。另一方面,KD-SGNN提出了两种KD机制(即软目标(ST)蒸馏和浅层模仿(SI)蒸馏)来提高表现力。KD-SGNN的可扩展性和有效性在多个真实数据集上进行了评估。此外,还通过综合分析验证了所提出的KD机制的有效性。