School of Software, Yunnan University, Kunming 650500, China.
Neural Netw. 2023 Oct;167:223-232. doi: 10.1016/j.neunet.2023.08.039. Epub 2023 Aug 24.
Graph representation is a critical technology in the field of knowledge engineering and knowledge-based applications since most knowledge bases are represented in the graph structure. Nowadays, contrastive learning has become a prominent way for graph representation by contrasting positive-positive and positive-negative node pairs between two augmentation graphs. It has achieved new state-of-the-art in the field of self-supervised graph representation. However, existing contrastive graph representation methods mainly focus on modifying (normally removing some edges/nodes) the original graph structure to generate the augmentation graph for the contrastive. It inevitably changes the original graph structures, meaning the generated augmentation graph is no longer equivalent to the original graph. This harms the performance of the representation in many structure-sensitive graphs such as protein graphs, chemical graphs, molecular graphs, etc. Moreover, there is only one positive-positive node pair but relatively massive positive-negative node pairs in the self-supervised graph contrastive learning. This can lead to the same class, or very similar samples are considered negative samples. To this end, in this work, we propose a Virtual Masking Augmentation (VMA) to generate an augmentation graph without changing any structures from the original graph. Meanwhile, a node augmentation method is proposed to augment the positive node pairs by discovering the most similar nodes in the same graph. Then, two different augmentation graphs are generated and put into a contrastive learning model to learn the graph representation. Extensive experiments on massive datasets demonstrate that our method achieves new state-of-the-art results on self-supervised graph representation. The source code of the proposed method is available at https://github.com/DuanhaoranCC/CGRA.
图表示是知识工程和基于知识的应用领域的关键技术,因为大多数知识库都以图结构表示。如今,对比学习已成为图表示的一种突出方式,通过对比两个增强图中的正-正和正-负节点对。它在自监督图表示领域取得了新的最先进水平。然而,现有的对比图表示方法主要集中在修改(通常是删除一些边/节点)原始图结构以生成对比的增强图。这不可避免地改变了原始的图结构,意味着生成的增强图不再等同于原始图。这在许多结构敏感的图中损害了表示的性能,例如蛋白质图、化学图、分子图等。此外,在自监督图对比学习中只有一个正-正节点对,但相对大量的正-负节点对。这可能导致同一类或非常相似的样本被视为负样本。为此,在这项工作中,我们提出了一种虚拟掩蔽增强(VMA)方法,从原始图中不改变任何结构来生成增强图。同时,提出了一种节点增强方法,通过发现同一图中最相似的节点来增强正节点对。然后,生成两个不同的增强图,并将它们放入对比学习模型中学习图表示。在大量数据集上的广泛实验表明,我们的方法在自监督图表示方面取得了新的最先进的结果。该方法的源代码可在 https://github.com/DuanhaoranCC/CGRA 上获得。