Chen Xueyuan, Li Shangzhe
State Key Laboratory of Software Development Environment, Beihang University, Beijing 100191, China.
School of Statistics and Mathematics, Central University of Finance and Economics, Beijing 100081, China.
Entropy (Basel). 2024 Feb 27;26(3):208. doi: 10.3390/e26030208.
Due to the success observed in deep neural networks with contrastive learning, there has been a notable surge in research interest in graph contrastive learning, primarily attributed to its superior performance in graphs with limited labeled data. Within contrastive learning, the selection of a "view" dictates the information captured by the representation, thereby influencing the model's performance. However, assessing the quality of information in these views poses challenges, and determining what constitutes a good view remains unclear. This paper addresses this issue by establishing the definition of a good view through the application of graph information bottleneck and structural entropy theories. Based on theoretical insights, we introduce CtrlGCL, a novel method for achieving a beneficial view in graph contrastive learning through coding tree representation learning. Extensive experiments were conducted to ascertain the effectiveness of the proposed view in unsupervised and semi-supervised learning. In particular, our approach, via CtrlGCL-H, yields an average accuracy enhancement of 1.06% under unsupervised learning when compared to GCL. This improvement underscores the efficacy of our proposed method.
由于在采用对比学习的深度神经网络中观察到的成功,对图对比学习的研究兴趣显著激增,这主要归因于其在标记数据有限的图中表现优异。在对比学习中,“视图”的选择决定了表示所捕获的信息,从而影响模型的性能。然而,评估这些视图中的信息质量具有挑战性,并且确定什么构成一个好的视图仍然不明确。本文通过应用图信息瓶颈和结构熵理论来定义一个好的视图,从而解决了这个问题。基于理论见解,我们引入了CtrlGCL,这是一种通过编码树表示学习在图对比学习中实现有益视图的新方法。进行了广泛的实验以确定所提出的视图在无监督和半监督学习中的有效性。特别是,与GCL相比,我们的方法通过CtrlGCL-H在无监督学习下平均准确率提高了1.06%。这一改进突出了我们所提出方法的有效性。