• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于聚类 Pareto 强化学习的多目标分子生成。

Multi-objective molecular generation via clustered Pareto-based reinforcement learning.

机构信息

School of Computer Science and Technology, Soochow University, Suzhou, 215006, China.

出版信息

Neural Netw. 2024 Nov;179:106596. doi: 10.1016/j.neunet.2024.106596. Epub 2024 Aug 6.

DOI:10.1016/j.neunet.2024.106596
PMID:39163823
Abstract

De novo molecular design is the process of learning knowledge from existing data to propose new chemical structures that satisfy the desired properties. By using de novo design to generate compounds in a directed manner, better solutions can be obtained in large chemical libraries with less comparison cost. But drug design needs to take multiple factors into consideration. For example, in polypharmacology, molecules that activate or inhibit multiple target proteins produce multiple pharmacological activities and are less susceptible to drug resistance. However, most existing molecular generation methods either focus only on affinity for a single target or fail to effectively balance the relationship between multiple targets, resulting in insufficient validity and desirability of the generated molecules. To address the problems, an approach called clustered Pareto-based reinforcement learning (CPRL) is proposed. In CPRL, a pre-trained model is constructed to grasp existing molecular knowledge in a supervised learning manner. In addition, the clustered Pareto optimization algorithm is presented to find the best solution between different objectives. The algorithm first extracts an update set from the sampled molecules through the designed aggregation-based molecular clustering. Then, the final reward is computed by constructing the Pareto frontier ranking of the molecules from the updated set. To explore the vast chemical space, a reinforcement learning agent is designed in CPRL that can be updated under the guidance of the final reward to balance multiple properties. Furthermore, to increase the internal diversity of the molecules, a fixed-parameter exploration model is used for sampling in conjunction with the agent. The experimental results demonstrate that CPRL is capable of balancing multiple properties of the molecule and has higher desirability and validity, reaching 0.9551 and 0.9923, respectively.

摘要

从头分子设计是指从现有数据中学习知识,提出满足所需性质的新化学结构的过程。通过使用从头设计以有方向的方式生成化合物,可以在具有较小比较成本的大型化学库中获得更好的解决方案。但是药物设计需要考虑多个因素。例如,在多药理学中,激活或抑制多个靶蛋白的分子会产生多种药理活性,并且不易产生抗药性。然而,大多数现有的分子生成方法要么只关注对单个靶标的亲和力,要么无法有效地平衡多个靶标之间的关系,导致生成分子的有效性和可取性不足。为了解决这些问题,提出了一种称为基于聚类 Pareto 的强化学习 (CPRL) 的方法。在 CPRL 中,构建了一个经过预训练的模型,以通过监督学习方式掌握现有的分子知识。此外,提出了聚类 Pareto 优化算法来在不同目标之间找到最佳解决方案。该算法首先通过设计的基于聚合的分子聚类从采样分子中提取更新集。然后,通过构建来自更新集的分子的 Pareto 前沿排名来计算最终奖励。为了探索广阔的化学空间,CPRL 中设计了一个强化学习代理,该代理可以在最终奖励的指导下进行更新,以平衡多个属性。此外,为了增加分子的内部多样性,使用固定参数探索模型与代理一起进行采样。实验结果表明,CPRL 能够平衡分子的多个属性,并且具有更高的可取性和有效性,分别达到 0.9551 和 0.9923。

相似文献

1
Multi-objective molecular generation via clustered Pareto-based reinforcement learning.基于聚类 Pareto 强化学习的多目标分子生成。
Neural Netw. 2024 Nov;179:106596. doi: 10.1016/j.neunet.2024.106596. Epub 2024 Aug 6.
2
DrugEx v2: de novo design of drug molecules by Pareto-based multi-objective reinforcement learning in polypharmacology.DrugEx v2:基于帕累托的多目标强化学习在多药理学中从头设计药物分子
J Cheminform. 2021 Nov 12;13(1):85. doi: 10.1186/s13321-021-00561-9.
3
FSM-DDTR: End-to-end feedback strategy for multi-objective De Novo drug design using transformers.FSM-DDTR:使用变压器的多目标从头药物设计的端到端反馈策略。
Comput Biol Med. 2023 Sep;164:107285. doi: 10.1016/j.compbiomed.2023.107285. Epub 2023 Jul 31.
4
De novo drug design based on Stack-RNN with multi-objective reward-weighted sum and reinforcement learning.基于堆叠循环神经网络的多目标奖励加权和强化学习的从头药物设计。
J Mol Model. 2023 Mar 30;29(4):121. doi: 10.1007/s00894-023-05523-6.
5
De novo drug design as GPT language modeling: large chemistry models with supervised and reinforcement learning.从头开始的药物设计与 GPT 语言模型:具有监督和强化学习的大型化学模型。
J Comput Aided Mol Des. 2024 Apr 22;38(1):20. doi: 10.1007/s10822-024-00559-z.
6
Faster and more diverse de novo molecular optimization with double-loop reinforcement learning using augmented SMILES.使用增强型 SMILES 进行双环强化学习,实现更快、更多样的从头分子优化。
J Comput Aided Mol Des. 2023 Aug;37(8):373-394. doi: 10.1007/s10822-023-00512-6. Epub 2023 Jun 17.
7
Computer-aided multi-objective optimization in small molecule discovery.小分子发现中的计算机辅助多目标优化
Patterns (N Y). 2023 Feb 10;4(2):100678. doi: 10.1016/j.patter.2023.100678.
8
Deep reinforcement learning for de novo drug design.基于深度强化学习的从头药物设计。
Sci Adv. 2018 Jul 25;4(7):eaap7885. doi: 10.1126/sciadv.aap7885. eCollection 2018 Jul.
9
Orientation-Preserving Rewards' Balancing in Reinforcement Learning.保持朝向的奖励在强化学习中的平衡。
IEEE Trans Neural Netw Learn Syst. 2022 Nov;33(11):6458-6472. doi: 10.1109/TNNLS.2021.3080521. Epub 2022 Oct 27.
10
De novo drug design by iterative multiobjective deep reinforcement learning with graph-based molecular quality assessment.基于图的分子质量评估的迭代多目标深度强化学习的从头药物设计。
Bioinformatics. 2023 Apr 3;39(4). doi: 10.1093/bioinformatics/btad157.