Suppr超能文献

少即是多:预训练图神经网络的数据主动视角

Better with Less: A Data-Active Perspective on Pre-Training Graph Neural Networks.

作者信息

Xu Jiarong, Huang Renhong, Jiang Xin, Cao Yuxuan, Yang Carl, Wang Chunping, Yang Yang

机构信息

Fudan University.

Zhejiang University.

出版信息

Adv Neural Inf Process Syst. 2023 Dec;36:56946-56978. Epub 2024 May 30.

Abstract

Pre-training on graph neural networks (GNNs) aims to learn transferable knowledge for downstream tasks with unlabeled data, and it has recently become an active research area. The success of graph pre-training models is often attributed to the massive amount of input data. In this paper, however, we identify the phenomenon in graph pre-training: more training data do not necessarily lead to better downstream performance. Motivated by this observation, we propose a framework for graph pre-training: fewer, but carefully chosen data are fed into a GNN model to enhance pre-training. The proposed pre-training pipeline is called the data-active graph pre-training (APT) framework, and is composed of a graph selector and a pre-training model. The graph selector chooses the most representative and instructive data points based on the inherent properties of graphs as well as . The proposed predictive uncertainty, as feedback from the pre-training model, measures the confidence level of the model in the data. When fed with the chosen data, on the other hand, the pre-training model grasps an initial understanding of the new, unseen data, and at the same time attempts to remember the knowledge learned from previous data. Therefore, the integration and interaction between these two components form a unified framework (APT), in which graph pre-training is performed in a progressive and iterative way. Experiment results show that the proposed APT is able to obtain an efficient pre-training model with fewer training data and better downstream performance.

摘要

图神经网络(GNN)的预训练旨在利用未标记数据学习可迁移知识,用于下游任务,并且最近已成为一个活跃的研究领域。图预训练模型的成功通常归因于大量的输入数据。然而,在本文中,我们发现了图预训练中的一种现象:更多的训练数据并不一定能带来更好的下游性能。受此观察结果的启发,我们提出了一种图预训练框架:将更少但经过精心挑选的数据输入到GNN模型中,以增强预训练效果。所提出的预训练流程称为数据主动图预训练(APT)框架,它由一个图选择器和一个预训练模型组成。图选择器根据图的固有属性以及所提出的预测不确定性(作为来自预训练模型的反馈,用于衡量模型对数据的置信度)来选择最具代表性和启发性的数据点。另一方面,当输入所选数据时,预训练模型对新的、未见过的数据形成初步理解,同时尝试记住从先前数据中学到的知识。因此,这两个组件之间的整合与交互形成了一个统一的框架(APT),其中图预训练以渐进和迭代的方式进行。实验结果表明,所提出的APT能够用更少的训练数据获得高效的预训练模型,并具有更好的下游性能。

相似文献

5
SP-GNN: Learning structure and position information from graphs.SP-GNN:从图中学习结构和位置信息。
Neural Netw. 2023 Apr;161:505-514. doi: 10.1016/j.neunet.2023.01.051. Epub 2023 Feb 4.
6
Semisupervised Graph Neural Networks for Graph Classification.用于图分类的半监督图神经网络
IEEE Trans Cybern. 2023 Oct;53(10):6222-6235. doi: 10.1109/TCYB.2022.3164696. Epub 2023 Sep 15.
9
Generalizing Graph Neural Networks on Out-of-Distribution Graphs.将图神经网络推广到分布外的图上。
IEEE Trans Pattern Anal Mach Intell. 2024 Jan;46(1):322-337. doi: 10.1109/TPAMI.2023.3321097. Epub 2023 Dec 5.
10
Auto-GNN: Neural architecture search of graph neural networks.自动图神经网络:图神经网络的神经架构搜索
Front Big Data. 2022 Nov 17;5:1029307. doi: 10.3389/fdata.2022.1029307. eCollection 2022.

本文引用的文献

5
MoleculeNet: a benchmark for molecular machine learning.分子网络:分子机器学习的一个基准
Chem Sci. 2017 Oct 31;9(2):513-530. doi: 10.1039/c7sc02664a. eCollection 2018 Jan 14.
6
Overcoming catastrophic forgetting in neural networks.克服神经网络中的灾难性遗忘。
Proc Natl Acad Sci U S A. 2017 Mar 28;114(13):3521-3526. doi: 10.1073/pnas.1611835114. Epub 2017 Mar 14.
8
Localization of the maximal entropy random walk.最大熵随机游走的定位
Phys Rev Lett. 2009 Apr 24;102(16):160602. doi: 10.1103/PhysRevLett.102.160602. Epub 2009 Apr 23.
9
Entropy rate of diffusion processes on complex networks.复杂网络上扩散过程的熵率
Phys Rev E Stat Nonlin Soft Matter Phys. 2008 Dec;78(6 Pt 2):065102. doi: 10.1103/PhysRevE.78.065102. Epub 2008 Dec 11.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验