• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

RetroCaptioner:通过对比标注的可学习图表示,在端到端逆合成变换器中超越注意力。

RetroCaptioner: beyond attention in end-to-end retrosynthesis transformer via contrastively captioned learnable graph representation.

机构信息

School of Chinese Materia Medica, Beijing University of Chinese Medicine, Beijing, 102488, China.

Ministry of Education, Engineering Research Center for Pharmaceutics of Chinese Materia Medica and New Drug Development, Beijing, 100102, China.

出版信息

Bioinformatics. 2024 Sep 2;40(9). doi: 10.1093/bioinformatics/btae561.

DOI:10.1093/bioinformatics/btae561
PMID:39342389
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11520410/
Abstract

MOTIVATION

Retrosynthesis identifies available precursor molecules for various and novel compounds. With the advancements and practicality of language models, Transformer-based models have increasingly been used to automate this process. However, many existing methods struggle to efficiently capture reaction transformation information, limiting the accuracy and applicability of their predictions.

RESULTS

We introduce RetroCaptioner, an advanced end-to-end, Transformer-based framework featuring a Contrastive Reaction Center Captioner. This captioner guides the training of dual-view attention models using a contrastive learning approach. It leverages learned molecular graph representations to capture chemically plausible constraints within a single-step learning process. We integrate the single-encoder, dual-encoder, and encoder-decoder paradigms to effectively fuse information from the sequence and graph representations of molecules. This involves modifying the Transformer encoder into a uni-view sequence encoder and a dual-view module. Furthermore, we enhance the captioning of atomic correspondence between SMILES and graphs. Our proposed method, RetroCaptioner, achieved outstanding performance with 67.2% in top-1 and 93.4% in top-10 exact matched accuracy on the USPTO-50k dataset, alongside an exceptional SMILES validity score of 99.4%. In addition, RetroCaptioner has demonstrated its reliability in generating synthetic routes for the drug protokylol.

AVAILABILITY AND IMPLEMENTATION

The code and data are available at https://github.com/guofei-tju/RetroCaptioner.

摘要

动机

逆合成分析可识别各种新颖化合物的可用前体分子。随着语言模型的进步和实用性的提高,基于转换器的模型越来越多地被用于自动化这一过程。然而,许多现有的方法难以有效地捕捉反应转化信息,限制了它们预测的准确性和适用性。

结果

我们引入了 RetroCaptioner,这是一种基于端到端的先进的基于转换器的框架,具有对比反应中心字幕器。该字幕器使用对比学习方法指导双视图注意力模型的训练。它利用学习到的分子图表示来在单个学习过程中捕获化学上合理的约束。我们集成了单编码器、双编码器和编码器-解码器范式,以有效地融合分子序列和图形表示的信息。这涉及将 Transformer 编码器修改为单视图序列编码器和双视图模块。此外,我们增强了 SMILES 和图形之间原子对应关系的字幕。我们提出的方法 RetroCaptioner 在 USPTO-50k 数据集上取得了卓越的性能,在 top-1 中达到了 67.2%的准确率,在 top-10 中达到了 93.4%的准确率,同时 SMILES 有效性得分达到了 99.4%。此外,RetroCaptioner 还证明了其在生成药物 protokylol 合成路线方面的可靠性。

可用性和实现

代码和数据可在 https://github.com/guofei-tju/RetroCaptioner 上获得。

相似文献

1
RetroCaptioner: beyond attention in end-to-end retrosynthesis transformer via contrastively captioned learnable graph representation.RetroCaptioner:通过对比标注的可学习图表示,在端到端逆合成变换器中超越注意力。
Bioinformatics. 2024 Sep 2;40(9). doi: 10.1093/bioinformatics/btae561.
2
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
3
Trajectory-Ordered Objectives for Self-Supervised Representation Learning of Temporal Healthcare Data Using Transformers: Model Development and Evaluation Study.使用Transformer进行时间序列医疗数据自监督表示学习的轨迹有序目标:模型开发与评估研究
JMIR Med Inform. 2025 Jun 4;13:e68138. doi: 10.2196/68138.
4
Short-Term Memory Impairment短期记忆障碍
5
Distilling knowledge from graph neural networks trained on cell graphs to non-neural student models.从在细胞图上训练的图神经网络中提取知识,用于非神经学生模型。
Sci Rep. 2025 Aug 10;15(1):29274. doi: 10.1038/s41598-025-13697-7.
6
Cognitive decline assessment using semantic linguistic content and transformer deep learning architecture.使用语义语言内容和变压器深度学习架构评估认知能力下降。
Int J Lang Commun Disord. 2024 May-Jun;59(3):1110-1127. doi: 10.1111/1460-6984.12973. Epub 2023 Nov 16.
7
Long-term care plan recommendation for older adults with disabilities: a bipartite graph transformer and self-supervised approach.针对残疾老年人的长期护理计划建议:一种二分图变压器和自监督方法。
J Am Med Inform Assoc. 2025 Apr 1;32(4):689-701. doi: 10.1093/jamia/ocae327.
8
HEART: Learning better representation of EHR data with a heterogeneous relation-aware transformer.心脏:使用异构关系感知转换器学习更好的 EHR 数据表示。
J Biomed Inform. 2024 Nov;159:104741. doi: 10.1016/j.jbi.2024.104741. Epub 2024 Oct 29.
9
A sequential recommendation method using contrastive learning and Wasserstein self-attention mechanism.一种使用对比学习和瓦瑟斯坦自注意力机制的序列推荐方法。
PeerJ Comput Sci. 2025 Mar 26;11:e2749. doi: 10.7717/peerj-cs.2749. eCollection 2025.
10
stGNN: Spatially Informed Cell-Type Deconvolution Based on Deep Graph Learning and Statistical Modeling.stGNN:基于深度图学习和统计建模的空间信息细胞类型反卷积
Interdiscip Sci. 2025 Jun 26. doi: 10.1007/s12539-025-00728-0.

引用本文的文献

1
AOPxSVM: A Support Vector Machine for Identifying Antioxidant Peptides Using a Block Substitution Matrix and Amino Acid Composition, Transformation, and Distribution Embeddings.AOPxSVM:一种使用块替换矩阵以及氨基酸组成、转化和分布嵌入来识别抗氧化肽的支持向量机。
Foods. 2025 Jun 6;14(12):2014. doi: 10.3390/foods14122014.

本文引用的文献

1
READRetro: natural product biosynthesis predicting with retrieval-augmented dual-view retrosynthesis.READRetro:基于检索增强的双视图回溯合成预测天然产物生物合成。
New Phytol. 2024 Sep;243(6):2512-2527. doi: 10.1111/nph.20012. Epub 2024 Jul 30.
2
MARS: a motif-based autoregressive model for retrosynthesis prediction.MARS:一种基于基序的回溯合成预测自回归模型。
Bioinformatics. 2024 Mar 4;40(3). doi: 10.1093/bioinformatics/btae115.
3
Retrosynthesis prediction with an interpretable deep-learning framework based on molecular assembly tasks.
基于分子组装任务的可解释深度学习框架进行逆合成预测。
Nat Commun. 2023 Oct 3;14(1):6155. doi: 10.1038/s41467-023-41698-5.
4
GRetro as a two-step graph generative models for retrosynthesis prediction.GRetro作为一种用于逆合成预测的两步图生成模型。
Commun Chem. 2023 May 30;6(1):102. doi: 10.1038/s42004-023-00897-3.
5
Retrosynthesis prediction using an end-to-end graph generative architecture for molecular graph editing.基于端到端图生成架构的分子图编辑回溯合成预测。
Nat Commun. 2023 May 25;14(1):3009. doi: 10.1038/s41467-023-38851-5.
6
G2GT: Retrosynthesis Prediction with Graph-to-Graph Attention Neural Network and Self-Training.G2GT:基于图到图注意力神经网络和自训练的逆合成预测
J Chem Inf Model. 2023 Apr 10;63(7):1894-1905. doi: 10.1021/acs.jcim.2c01302. Epub 2023 Mar 22.
7
Merging enzymatic and synthetic chemistry with computational synthesis planning.将酶学和合成化学与计算合成规划相结合。
Nat Commun. 2022 Dec 14;13(1):7747. doi: 10.1038/s41467-022-35422-y.
8
Root-aligned SMILES: a tight representation for chemical reaction prediction.根对齐的简化分子线性输入规范(SMILES):一种用于化学反应预测的紧凑表示法。
Chem Sci. 2022 Jul 12;13(31):9023-9034. doi: 10.1039/d2sc02763a. eCollection 2022 Aug 10.
9
Permutation Invariant Graph-to-Sequence Model for Template-Free Retrosynthesis and Reaction Prediction.无模板回溯合成和反应预测的置换不变图到序列模型。
J Chem Inf Model. 2022 Aug 8;62(15):3503-3513. doi: 10.1021/acs.jcim.2c00321. Epub 2022 Jul 26.
10
Deep learning driven biosynthetic pathways navigation for natural products with BioNavi-NP.深度学习驱动的生物合成途径导航用于天然产物的 BioNavi-NP。
Nat Commun. 2022 Jun 10;13(1):3342. doi: 10.1038/s41467-022-30970-9.