• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

阿里尔:对抗性图对比学习

ArieL: Adversarial Graph Contrastive Learning.

作者信息

Feng Shengyu, Jing Baoyu, Zhu Yada, Tong Hanghang

机构信息

Carnegie Mellon University, USA.

University of Illinois at Urbana-Champaign, USA.

出版信息

ACM Trans Knowl Discov Data. 2023 Dec;18(4):1-22. doi: 10.1145/3638054.

DOI:10.1145/3638054
PMID:40264574
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12012824/
Abstract

Contrastive learning is an effective unsupervised method in graph representation learning, and the key component of contrastive learning lies in the construction of positive and negative samples. Previous methods usually utilize the proximity of nodes in the graph as the principle. Recently, the data-augmentation-based contrastive learning method has advanced to show great power in the visual domain, and some works extended this method from images to graphs. However, unlike the data augmentation on images, the data augmentation on graphs is far less intuitive and much harder to provide high-quality contrastive samples, which leaves much space for improvement. In this work, by introducing an adversarial graph view for data augmentation, we propose a simple but effective method, (ArieL), to extract informative contrastive samples within reasonable constraints. We develop a new technique called information regularization for stable training and use subgraph sampling for scalability. We generalize our method from node-level contrastive learning to the graph level by treating each graph instance as a super-node. ArieL consistently outperforms the current graph contrastive learning methods for both node-level and graph-level classification tasks on real-world datasets. We further demonstrate that ArieL is more robust in the face of adversarial attacks.

摘要

对比学习是图表示学习中一种有效的无监督方法,对比学习的关键在于正负样本的构建。以往的方法通常以图中节点的邻近性为原则。最近,基于数据增强的对比学习方法在视觉领域取得了进展,并显示出强大的能力,一些工作将该方法从图像扩展到了图。然而,与图像上的数据增强不同,图上的数据增强远没有那么直观,而且很难提供高质量的对比样本,这就留下了很大的改进空间。在这项工作中,通过引入用于数据增强的对抗性图视图,我们提出了一种简单而有效的方法(ArieL),以在合理的约束内提取信息丰富的对比样本。我们开发了一种名为信息正则化的新技术用于稳定训练,并使用子图采样来提高可扩展性。通过将每个图实例视为一个超级节点,我们将我们的方法从节点级对比学习推广到图级。在真实世界数据集上的节点级和图级分类任务中,ArieL始终优于当前的图对比学习方法。我们进一步证明,ArieL在面对对抗性攻击时更具鲁棒性。

相似文献

1
ArieL: Adversarial Graph Contrastive Learning.阿里尔:对抗性图对比学习
ACM Trans Knowl Discov Data. 2023 Dec;18(4):1-22. doi: 10.1145/3638054.
2
Multiscale Subgraph Adversarial Contrastive Learning.多尺度子图对抗对比学习
IEEE Trans Neural Netw Learn Syst. 2025 Aug;36(8):15001-15014. doi: 10.1109/TNNLS.2025.3543954.
3
ERMAV: Efficient and Robust Graph Contrastive Learning via Multiadversarial Views Training.ERMAV:通过多对抗视图训练实现高效且稳健的图对比学习
IEEE Trans Cybern. 2025 May;55(5):2188-2201. doi: 10.1109/TCYB.2025.3548175. Epub 2025 Apr 23.
4
Contrastive Graph Representation Learning with Adversarial Cross-View Reconstruction and Information Bottleneck.基于对抗性跨视图重建和信息瓶颈的对比图表示学习
Neural Netw. 2025 Apr;184:107094. doi: 10.1016/j.neunet.2024.107094. Epub 2025 Jan 9.
5
Self-supervised contrastive graph representation with node and graph augmentation.自监督对比图表示与节点和图增强。
Neural Netw. 2023 Oct;167:223-232. doi: 10.1016/j.neunet.2023.08.039. Epub 2023 Aug 24.
6
GMNI: Achieve good data augmentation in unsupervised graph contrastive learning.GMNI:在无监督图对比学习中实现良好的数据增强。
Neural Netw. 2025 Jan;181:106804. doi: 10.1016/j.neunet.2024.106804. Epub 2024 Oct 18.
7
Contrastive learning of graphs under label noise.图在标签噪声下的对比学习。
Neural Netw. 2024 Apr;172:106113. doi: 10.1016/j.neunet.2024.106113. Epub 2024 Jan 6.
8
Accurate graph classification via two-staged contrastive curriculum learning.通过两阶段对比课程学习实现准确的图分类。
PLoS One. 2024 Jan 3;19(1):e0296171. doi: 10.1371/journal.pone.0296171. eCollection 2024.
9
Graph contrastive learning with node-level accurate difference.具有节点级精确差异的图对比学习
Fundam Res. 2024 Sep 3;5(2):818-829. doi: 10.1016/j.fmre.2024.06.013. eCollection 2025 Mar.
10
A Topology-Enhanced Multi-Viewed Contrastive Approach for Molecular Graph Representation Learning and Classification.一种用于分子图表示学习和分类的拓扑增强多视图对比方法。
Mol Inform. 2025 Jan;44(1):e202400252. doi: 10.1002/minf.202400252.

引用本文的文献

1
Deep clustering of single-cell RNA-seq using adversarial graph contrastive learning.使用对抗性图对比学习对单细胞RNA测序进行深度聚类。
Brief Bioinform. 2025 Jul 2;26(4). doi: 10.1093/bib/bbaf423.
2
Dual-level graph contrastive collaborative filtering.双层次图对比协同过滤
Sci Rep. 2025 Jul 9;15(1):24734. doi: 10.1038/s41598-025-10621-x.
3
Drug-drug interaction analysis based on information bottleneck graph neural network: A review.基于信息瓶颈图神经网络的药物-药物相互作用分析:综述
Medicine (Baltimore). 2025 Jun 20;104(25):e42904. doi: 10.1097/MD.0000000000042904.

本文引用的文献

1
node2vec: Scalable Feature Learning for Networks.节点2向量:网络的可扩展特征学习
KDD. 2016 Aug;2016:855-864. doi: 10.1145/2939672.2939754.