Suppr超能文献

将图神经网络推广到分布外的图上。

Generalizing Graph Neural Networks on Out-of-Distribution Graphs.

作者信息

Fan Shaohua, Wang Xiao, Shi Chuan, Cui Peng, Wang Bai

出版信息

IEEE Trans Pattern Anal Mach Intell. 2024 Jan;46(1):322-337. doi: 10.1109/TPAMI.2023.3321097. Epub 2023 Dec 5.

Abstract

Graph Neural Networks (GNNs) are proposed without considering the agnostic distribution shifts between training graphs and testing graphs, inducing the degeneration of the generalization ability of GNNs in Out-Of-Distribution (OOD) settings. The fundamental reason for such degeneration is that most GNNs are developed based on the I.I.D hypothesis. In such a setting, GNNs tend to exploit subtle statistical correlations existing in the training set for predictions, even though it is a spurious correlation. This learning mechanism inherits from the common characteristics of machine learning approaches. However, such spurious correlations may change in the wild testing environments, leading to the failure of GNNs. Therefore, eliminating the impact of spurious correlations is crucial for stable GNN models. To this end, in this paper, we argue that the spurious correlation exists among subgraph-level units and analyze the degeneration of GNN in causal view. Based on the causal view analysis, we propose a general causal representation framework for stable GNN, called StableGNN. The main idea of this framework is to extract high-level representations from raw graph data first and resort to the distinguishing ability of causal inference to help the model get rid of spurious correlations. Particularly, to extract meaningful high-level representations, we exploit a differentiable graph pooling layer to extract subgraph-based representations by an end-to-end manner. Furthermore, inspired by the confounder balancing techniques from causal inference, based on the learned high-level representations, we propose a causal variable distinguishing regularizer to correct the biased training distribution by learning a set of sample weights. Hence, GNNs would concentrate more on the true connection between discriminative substructures and labels. Extensive experiments are conducted on both synthetic datasets with various distribution shift degrees and eight real-world OOD graph datasets. The results well verify that the proposed model StableGNN not only outperforms the state-of-the-arts but also provides a flexible framework to enhance existing GNNs. In addition, the interpretability experiments validate that StableGNN could leverage causal structures for predictions.

摘要

图神经网络(GNN)在提出时未考虑训练图和测试图之间不可知的分布偏移,这导致GNN在分布外(OOD)设置下泛化能力的退化。这种退化的根本原因是大多数GNN是基于独立同分布(I.I.D)假设开发的。在这种情况下,GNN倾向于利用训练集中存在的微妙统计相关性进行预测,即使这是一种虚假相关性。这种学习机制继承了机器学习方法的共同特征。然而,这种虚假相关性在实际测试环境中可能会发生变化,从而导致GNN失效。因此,消除虚假相关性的影响对于稳定的GNN模型至关重要。为此,在本文中,我们认为虚假相关性存在于子图级单元之间,并从因果视角分析了GNN的退化情况。基于因果视角分析,我们提出了一种用于稳定GNN的通用因果表示框架,称为StableGNN。该框架的主要思想是首先从原始图数据中提取高级表示,并借助因果推理的区分能力帮助模型摆脱虚假相关性。具体而言,为了提取有意义的高级表示,我们利用一个可微图池化层以端到端方式提取基于子图的表示。此外,受因果推理中的混杂因素平衡技术启发,基于学习到的高级表示,我们提出了一种因果变量区分正则化器,通过学习一组样本权重来校正有偏差的训练分布。因此,GNN将更多地关注判别性子结构和标签之间的真实联系。我们在具有不同分布偏移程度的合成数据集和八个真实世界的OOD图数据集上进行了广泛的实验。结果很好地验证了所提出的模型StableGNN不仅优于现有技术,还提供了一个灵活的框架来增强现有的GNN。此外,可解释性实验验证了StableGNN可以利用因果结构进行预测。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验