• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于图神经网络的原子体系注意力迁移学习(TAAG)。

Transfer learning using attentions across atomic systems with graph neural networks (TAAG).

机构信息

Department of Chemical Engineering, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA.

Meta AI Research, Menlo Park, California 94025, USA.

出版信息

J Chem Phys. 2022 May 14;156(18):184702. doi: 10.1063/5.0088019.

DOI:10.1063/5.0088019
PMID:35568535
Abstract

Recent advances in Graph Neural Networks (GNNs) have transformed the space of molecular and catalyst discovery. Despite the fact that the underlying physics across these domains remain the same, most prior work has focused on building domain-specific models either in small molecules or in materials. However, building large datasets across all domains is computationally expensive; therefore, the use of transfer learning (TL) to generalize to different domains is a promising but under-explored approach to this problem. To evaluate this hypothesis, we use a model that is pretrained on the Open Catalyst Dataset (OC20), and we study the model's behavior when fine-tuned for a set of different datasets and tasks. This includes MD17, the *CO adsorbate dataset, and OC20 across different tasks. Through extensive TL experiments, we demonstrate that the initial layers of GNNs learn a more basic representation that is consistent across domains, whereas the final layers learn more task-specific features. Moreover, these well-known strategies show significant improvement over the non-pretrained models for in-domain tasks with improvements of 53% and 17% for the *CO dataset and across the Open Catalyst Project (OCP) task, respectively. TL approaches result in up to 4× speedup in model training depending on the target data and task. However, these do not perform well for the MD17 dataset, resulting in worse performance than the non-pretrained model for few molecules. Based on these observations, we propose transfer learning using attentions across atomic systems with graph Neural Networks (TAAG), an attention-based approach that adapts to prioritize and transfer important features from the interaction layers of GNNs. The proposed method outperforms the best TL approach for out-of-domain datasets, such as MD17, and gives a mean improvement of 6% over a model trained from scratch.

摘要

最近图神经网络 (GNN) 的进展改变了分子和催化剂发现的领域。尽管这些领域的基础物理仍然相同,但大多数之前的工作都集中在小分子或材料领域构建特定于领域的模型上。然而,在所有领域构建大型数据集在计算上是昂贵的;因此,使用迁移学习 (TL) 将模型推广到不同的领域是解决这个问题的一种很有前途但尚未充分探索的方法。为了评估这个假设,我们使用在 Open Catalyst Dataset (OC20) 上预训练的模型,并研究了在一组不同的数据集和任务上进行微调时模型的行为。这包括 MD17、*CO 吸附物数据集和 OC20 的不同任务。通过广泛的 TL 实验,我们证明 GNN 的初始层学习了一种更基本的表示,这种表示在不同的领域是一致的,而最后一层学习了更多的任务特定特征。此外,这些广为人知的策略在域内任务上取得了显著的改进,与非预训练模型相比,*CO 数据集的改进分别为 53%和 17%,在 Open Catalyst Project (OCP) 任务上的改进分别为 53%和 17%。TL 方法在模型训练中最多可以实现 4 倍的速度提升,具体取决于目标数据和任务。然而,对于 MD17 数据集,这些方法的效果并不理想,对于少数分子,其性能甚至比非预训练模型还要差。基于这些观察,我们提出了使用注意力在原子系统之间进行图神经网络的迁移学习 (TAAG),这是一种基于注意力的方法,它可以适应从 GNN 的交互层中优先和转移重要特征。所提出的方法在 MD17 等域外数据集上优于最佳 TL 方法,并比从头开始训练的模型提高了 6%的平均水平。

相似文献

1
Transfer learning using attentions across atomic systems with graph neural networks (TAAG).基于图神经网络的原子体系注意力迁移学习(TAAG)。
J Chem Phys. 2022 May 14;156(18):184702. doi: 10.1063/5.0088019.
2
Targeted transfer learning to improve performance in small medical physics datasets.靶向迁移学习以提高小型医学物理数据集的性能。
Med Phys. 2020 Dec;47(12):6246-6256. doi: 10.1002/mp.14507. Epub 2020 Oct 25.
3
Enhancing molecular property prediction with auxiliary learning and task-specific adaptation.通过辅助学习和特定任务适应增强分子性质预测。
J Cheminform. 2024 Jul 24;16(1):85. doi: 10.1186/s13321-024-00880-7.
4
MGLNN: Semi-supervised learning via Multiple Graph Cooperative Learning Neural Networks.MGLNN:基于多图协同学习神经网络的半监督学习。
Neural Netw. 2022 Sep;153:204-214. doi: 10.1016/j.neunet.2022.05.024. Epub 2022 Jun 3.
5
Augmented Graph Neural Network with hierarchical global-based residual connections.基于层次全局残差连接的增强图神经网络。
Neural Netw. 2022 Jun;150:149-166. doi: 10.1016/j.neunet.2022.03.008. Epub 2022 Mar 10.
6
A transfer learning model with multi-source domains for biomedical event trigger extraction.一种用于生物医学事件触发词提取的多源域迁移学习模型。
BMC Genomics. 2021 Jan 7;22(1):31. doi: 10.1186/s12864-020-07315-1.
7
Finding core labels for maximizing generalization of graph neural networks.发现图神经网络泛化最大化的核心标签。
Neural Netw. 2024 Dec;180:106635. doi: 10.1016/j.neunet.2024.106635. Epub 2024 Aug 14.
8
A Comprehensive Survey on Graph Neural Networks.图神经网络综述。
IEEE Trans Neural Netw Learn Syst. 2021 Jan;32(1):4-24. doi: 10.1109/TNNLS.2020.2978386. Epub 2021 Jan 4.
9
AutoTune: Automatically Tuning Convolutional Neural Networks for Improved Transfer Learning.AutoTune:自动调整卷积神经网络以提高迁移学习性能。
Neural Netw. 2021 Jan;133:112-122. doi: 10.1016/j.neunet.2020.10.009. Epub 2020 Oct 27.
10
ReLMole: Molecular Representation Learning Based on Two-Level Graph Similarities.ReLMole:基于两级图相似度的分子表示学习
J Chem Inf Model. 2022 Nov 28;62(22):5361-5372. doi: 10.1021/acs.jcim.2c00798. Epub 2022 Oct 27.

引用本文的文献

1
DPA-2: a large atomic model as a multi-task learner.DPA - 2:作为多任务学习者的大型原子模型。
NPJ Comput Mater. 2024;10(1). doi: 10.1038/s41524-024-01493-2. Epub 2024 Dec 19.
2
Enhancing Machine Learning Potentials through Transfer Learning across Chemical Elements.通过跨化学元素的迁移学习提升机器学习潜力。
J Chem Inf Model. 2025 Jul 28;65(14):7406-7414. doi: 10.1021/acs.jcim.5c00293. Epub 2025 Jul 7.
3
Electrochemical CO reduction to liquid fuels: Mechanistic pathways and surface/interface engineering of catalysts and electrolytes.
电化学CO还原为液体燃料:催化剂和电解质的机理途径及表面/界面工程
Innovation (Camb). 2025 Jan 17;6(3):100807. doi: 10.1016/j.xinn.2025.100807. eCollection 2025 Mar 3.
4
From bulk effective mass to 2D carrier mobility accurate prediction via adversarial transfer learning.通过对抗性迁移学习实现从体有效质量到二维载流子迁移率的精确预测。
Nat Commun. 2024 Jun 25;15(1):5391. doi: 10.1038/s41467-024-49686-z.
5
Applying Large Graph Neural Networks to Predict Transition Metal Complex Energies Using the tmQM_wB97MV Data Set.应用大型图神经网络使用 tmQM_wB97MV 数据集预测过渡金属配合物能量。
J Chem Inf Model. 2023 Dec 25;63(24):7642-7654. doi: 10.1021/acs.jcim.3c01226. Epub 2023 Dec 4.