You Yuning, Chen Tianlong, Wang Zhangyang, Shen Yang
Texas A&M University.
University of Texas at Austin.
Proc Int Conf Web Search Data Min. 2022 Feb;2022:1300-1309. doi: 10.1145/3488560.3498416. Epub 2022 Feb 15.
Self-supervision is recently surging at its new frontier of graph learning. It facilitates graph representations beneficial to downstream tasks; but its success could hinge on domain knowledge for handcraft or the often expensive trials and errors. Even its state-of-the-art representative, graph contrastive learning (GraphCL), is not completely free of those needs as GraphCL uses a prefabricated prior reflected by the ad-hoc manual selection of graph data augmentations. Our work aims at advancing GraphCL by answering the following questions: Accordingly, we have extended the prefabricated discrete prior in the augmentation set, to a learnable continuous prior in the parameter space of graph generators, assuming that graph priors , similar to the concept of image manifolds, can be learned by data generation. Furthermore, to form contrastive views without collapsing to trivial solutions due to the prior learnability, we have leveraged both principles of information minimization (InfoMin) and information bottleneck (InfoBN) to regularize the learned priors. Eventually, contrastive learning, InfoMin, and InfoBN are incorporated organically into one framework of bi-level optimization. Our principled and automated approach has proven to be competitive against the state-of-the-art graph self-supervision methods, including GraphCL, on benchmarks of small graphs; and shown even better generalizability on large-scale graphs, without resorting to human expertise or downstream validation. Our code is publicly released at https://github.com/Shen-Lab/GraphCL_Automated.
自我监督最近在图学习的新前沿领域蓬勃发展。它有助于生成对下游任务有益的图表示;但其成功可能取决于手工制作的领域知识或通常代价高昂的反复试验。即使是其最先进的代表性方法——图对比学习(GraphCL),也并非完全摆脱了这些需求,因为GraphCL使用了由临时手动选择图数据增强所反映的预制先验。我们的工作旨在通过回答以下问题来推进GraphCL:因此,我们将增强集中预制的离散先验扩展为图生成器参数空间中可学习的连续先验,假设图先验类似于图像流形的概念,可以通过数据生成来学习。此外,为了在不由于先验可学习性而陷入平凡解的情况下形成对比视图,我们利用了信息最小化(InfoMin)和信息瓶颈(InfoBN)这两个原则来正则化学习到的先验。最终,对比学习、InfoMin和InfoBN被有机地整合到一个双层优化框架中。我们基于原则的自动化方法已被证明在小图基准测试中与包括GraphCL在内的最先进的图自我监督方法具有竞争力;并且在大规模图上显示出更好的泛化能力,而无需借助人类专业知识或下游验证。我们的代码已在https://github.com/Shen-Lab/GraphCL_Automated上公开发布。