Suppr超能文献

用于分子设计的基于主动学习的样本高效强化学习

Sample efficient reinforcement learning with active learning for molecular design.

作者信息

Dodds Michael, Guo Jeff, Löhr Thomas, Tibo Alessandro, Engkvist Ola, Janet Jon Paul

机构信息

Molecular AI, Discovery Sciences, R&D, AstraZeneca 431 50 Gothenburg Sweden

出版信息

Chem Sci. 2024 Feb 8;15(11):4146-4160. doi: 10.1039/d3sc04653b. eCollection 2024 Mar 13.

Abstract

Reinforcement learning (RL) is a powerful and flexible paradigm for searching for solutions in high-dimensional action spaces. However, bridging the gap between playing computer games with thousands of simulated episodes and solving real scientific problems with complex and involved environments (up to actual laboratory experiments) requires improvements in terms of sample efficiency to make the most of expensive information. The discovery of new drugs is a major commercial application of RL, motivated by the very large nature of the chemical space and the need to perform multiparameter optimization (MPO) across different properties. methods, such as virtual library screening (VS) and molecular generation with RL, show great promise in accelerating this search. However, incorporation of increasingly complex computational models in these workflows requires increasing sample efficiency. Here, we introduce an active learning system linked with an RL model (RL-AL) for molecular design, which aims to improve the sample-efficiency of the optimization process. We identity and characterize unique challenges combining RL and AL, investigate the interplay between the systems, and develop a novel AL approach to solve the MPO problem. Our approach greatly expedites the search for novel solutions relative to baseline-RL for simple ligand- and structure-based oracle functions, with a 5-66-fold increase in hits generated for a fixed oracle budget and a 4-64-fold reduction in computational time to find a specific number of hits. Furthermore, compounds discovered through RL-AL display substantial enrichment of a multi-parameter scoring objective, indicating superior efficacy in curating high-scoring compounds, without a reduction in output diversity. This significant acceleration improves the feasibility of oracle functions that have largely been overlooked in RL due to high computational costs, for example free energy perturbation methods, and in principle is applicable to any RL domain.

摘要

强化学习(RL)是一种在高维动作空间中寻找解决方案的强大且灵活的范式。然而,要弥合通过数千次模拟情节玩电脑游戏与在复杂且棘手的环境(直至实际实验室实验)中解决实际科学问题之间的差距,就需要在样本效率方面加以改进,以便充分利用昂贵的信息。新药发现是强化学习的一项主要商业应用,其动机源于化学空间的巨大规模以及对跨不同属性进行多参数优化(MPO)的需求。虚拟库筛选(VS)和利用强化学习进行分子生成等方法在加速这一搜索过程中显示出巨大潜力。然而,在这些工作流程中纳入日益复杂的计算模型需要提高样本效率。在此,我们引入了一种与强化学习模型(RL-AL)相关联的主动学习系统用于分子设计,旨在提高优化过程的样本效率。我们识别并刻画了将强化学习与主动学习相结合所面临的独特挑战,研究了各系统之间的相互作用,并开发了一种新颖的主动学习方法来解决多参数优化问题。相对于基于简单配体和结构的预言函数的基线强化学习,我们的方法极大地加快了寻找新解决方案的速度,在固定的预言预算下,命中次数增加了5至66倍,而找到特定数量命中结果的计算时间减少了4至64倍。此外,通过RL-AL发现的化合物在多参数评分目标上有显著富集,这表明在筛选高分化合物方面具有卓越功效,同时输出多样性并未降低。这种显著的加速提高了由于计算成本高而在强化学习中 largely 被忽视的预言函数的可行性,例如自由能微扰方法,并且原则上适用于任何强化学习领域。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e919/10935729/ecd963225897/d3sc04653b-f1.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验