• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Learning in continuous action space for developing high dimensional potential energy models.在连续动作空间中学习,以开发高维势能模型。
Nat Commun. 2022 Jan 18;13(1):368. doi: 10.1038/s41467-021-27849-6.
2
Active Inference and Reinforcement Learning: A Unified Inference on Continuous State and Action Spaces Under Partial Observability.主动推理与强化学习:部分可观测性下连续状态与动作空间的统一推理
Neural Comput. 2024 Sep 17;36(10):2073-2135. doi: 10.1162/neco_a_01698.
3
Folic acid supplementation and malaria susceptibility and severity among people taking antifolate antimalarial drugs in endemic areas.在流行地区,服用抗叶酸抗疟药物的人群中,叶酸补充剂与疟疾易感性和严重程度的关系。
Cochrane Database Syst Rev. 2022 Feb 1;2(2022):CD014217. doi: 10.1002/14651858.CD014217.
4
Sample efficient reinforcement learning with active learning for molecular design.用于分子设计的基于主动学习的样本高效强化学习
Chem Sci. 2024 Feb 8;15(11):4146-4160. doi: 10.1039/d3sc04653b. eCollection 2024 Mar 13.
5
Representation learning for continuous action spaces is beneficial for efficient policy learning.连续动作空间的表示学习有利于高效的策略学习。
Neural Netw. 2023 Feb;159:137-152. doi: 10.1016/j.neunet.2022.12.009. Epub 2022 Dec 16.
6
A Deep Reinforcement Learning-Based MPPT Control for PV Systems under Partial Shading Condition.一种基于深度强化学习的部分阴影条件下光伏系统最大功率点跟踪控制方法
Sensors (Basel). 2020 May 27;20(11):3039. doi: 10.3390/s20113039.
7
Relative Entropy Regularized Sample-Efficient Reinforcement Learning With Continuous Actions.具有连续动作的相对熵正则化样本高效强化学习
IEEE Trans Neural Netw Learn Syst. 2025 Jan;36(1):475-485. doi: 10.1109/TNNLS.2023.3329513. Epub 2025 Jan 7.
8
BNAS: Efficient Neural Architecture Search Using Broad Scalable Architecture.BNAS:使用广泛可扩展架构的高效神经架构搜索。
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):5004-5018. doi: 10.1109/TNNLS.2021.3067028. Epub 2022 Aug 31.
9
Exploration in neo-Hebbian reinforcement learning: Computational approaches to the exploration-exploitation balance with bio-inspired neural networks.神经拟态强化学习探索:基于生物启发神经网络的探索-利用平衡计算方法。
Neural Netw. 2022 Jul;151:16-33. doi: 10.1016/j.neunet.2022.03.021. Epub 2022 Mar 23.
10
Learning with Delayed Rewards-A Case Study on Inverse Defect Design in 2D Materials.基于延迟奖励的学习——二维材料逆缺陷设计的案例研究
ACS Appl Mater Interfaces. 2021 Aug 4;13(30):36455-36464. doi: 10.1021/acsami.1c07545. Epub 2021 Jul 21.

引用本文的文献

1
Comparing ANI-2x, ANI-1ccx neural networks, force field, and DFT methods for predicting conformational potential energy of organic molecules.比较ANI-2x、ANI-1ccx神经网络、力场和密度泛函理论(DFT)方法预测有机分子构象势能的能力。
Sci Rep. 2024 May 23;14(1):11791. doi: 10.1038/s41598-024-62242-5.
2
Employing neural density functionals to generate potential energy surfaces.利用神经密度泛函生成势能面。
J Mol Model. 2024 Feb 10;30(3):65. doi: 10.1007/s00894-024-05834-2.
3
Identifying the communication of burnout syndrome on the Twitter platform from the individual, organizational, and environmental perspective.从个人、组织和环境角度识别推特平台上职业倦怠综合征的传播情况。
Front Psychol. 2023 Oct 19;14:1236491. doi: 10.3389/fpsyg.2023.1236491. eCollection 2023.
4
Materials Discovery With Machine Learning and Knowledge Discovery.基于机器学习和知识发现的材料探索
Front Chem. 2022 Jul 7;10:930369. doi: 10.3389/fchem.2022.930369. eCollection 2022.
5
Machine learning the metastable phase diagram of covalently bonded carbon.通过机器学习确定共价键合碳的亚稳相图。
Nat Commun. 2022 Jun 6;13(1):3251. doi: 10.1038/s41467-022-30820-8.

本文引用的文献

1
Towards efficient discovery of green synthetic pathways with Monte Carlo tree search and reinforcement learning.利用蒙特卡罗树搜索和强化学习实现绿色合成途径的高效发现。
Chem Sci. 2020 Sep 14;11(40):10959-10972. doi: 10.1039/d0sc04184j.
2
Single atom catalysis poised to transition from an academic curiosity to an industrially relevant technology.单原子催化有望从学术好奇心转变为具有工业相关性的技术。
Nat Commun. 2021 Feb 9;12(1):895. doi: 10.1038/s41467-021-21152-0.
3
Microenvironment modulation of single-atom catalysts and their roles in electrochemical energy conversion.单原子催化剂的微环境调控及其在电化学能量转换中的作用
Sci Adv. 2020 Sep 23;6(39). doi: 10.1126/sciadv.abb6833. Print 2020 Sep.
4
A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play.一种通过自我对弈掌握国际象棋、将棋和围棋的通用强化学习算法。
Science. 2018 Dec 7;362(6419):1140-1144. doi: 10.1126/science.aar6404.
5
Characterization of Elastic Modulus Across the (AlSc)N System Using DFT and Substrate-Effect-Corrected Nanoindentation.使用密度泛函理论(DFT)和衬底效应校正纳米压痕对(AlSc)N系统的弹性模量进行表征
IEEE Trans Ultrason Ferroelectr Freq Control. 2018 Nov;65(11):2167-2175. doi: 10.1109/TUFFC.2018.2862240. Epub 2018 Aug 16.
6
Deep reinforcement learning for de novo drug design.基于深度强化学习的从头药物设计。
Sci Adv. 2018 Jul 25;4(7):eaap7885. doi: 10.1126/sciadv.aap7885. eCollection 2018 Jul.
7
Machine learning for molecular and materials science.机器学习在分子和材料科学中的应用。
Nature. 2018 Jul;559(7715):547-555. doi: 10.1038/s41586-018-0337-2. Epub 2018 Jul 25.
8
Atomically Precise Clusters of Noble Metals: Emerging Link between Atoms and Nanoparticles.原子级精确的贵金属原子簇:原子与纳米粒子间的新兴连接。
Chem Rev. 2017 Jun 28;117(12):8208-8271. doi: 10.1021/acs.chemrev.6b00769. Epub 2017 Jun 6.
9
Mastering the game of Go with deep neural networks and tree search.用深度神经网络和树搜索掌握围棋游戏。
Nature. 2016 Jan 28;529(7587):484-9. doi: 10.1038/nature16961.
10
The Birmingham parallel genetic algorithm and its application to the direct DFT global optimisation of Ir(N) (N = 10-20) clusters.伯明翰并行遗传算法及其在Ir(N)(N = 10 - 20)团簇直接离散傅里叶变换全局优化中的应用。
Nanoscale. 2015 Sep 7;7(33):14032-8. doi: 10.1039/c5nr03774c. Epub 2015 Aug 4.

在连续动作空间中学习,以开发高维势能模型。

Learning in continuous action space for developing high dimensional potential energy models.

机构信息

Center for Nanoscale Materials, Argonne National Laboratory, Lemont, IL, 60439, USA.

Department of Mechanical and Industrial Engineering, University of Illinois, Chicago, IL, 60607, USA.

出版信息

Nat Commun. 2022 Jan 18;13(1):368. doi: 10.1038/s41467-021-27849-6.

DOI:10.1038/s41467-021-27849-6
PMID:35042872
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8766468/
Abstract

Reinforcement learning (RL) approaches that combine a tree search with deep learning have found remarkable success in searching exorbitantly large, albeit discrete action spaces, as in chess, Shogi and Go. Many real-world materials discovery and design applications, however, involve multi-dimensional search problems and learning domains that have continuous action spaces. Exploring high-dimensional potential energy models of materials is an example. Traditionally, these searches are time consuming (often several years for a single bulk system) and driven by human intuition and/or expertise and more recently by global/local optimization searches that have issues with convergence and/or do not scale well with the search dimensionality. Here, in a departure from discrete action and other gradient-based approaches, we introduce a RL strategy based on decision trees that incorporates modified rewards for improved exploration, efficient sampling during playouts and a "window scaling scheme" for enhanced exploitation, to enable efficient and scalable search for continuous action space problems. Using high-dimensional artificial landscapes and control RL problems, we successfully benchmark our approach against popular global optimization schemes and state of the art policy gradient methods, respectively. We demonstrate its efficacy to parameterize potential models (physics based and high-dimensional neural networks) for 54 different elemental systems across the periodic table as well as alloys. We analyze error trends across different elements in the latent space and trace their origin to elemental structural diversity and the smoothness of the element energy surface. Broadly, our RL strategy will be applicable to many other physical science problems involving search over continuous action spaces.

摘要

强化学习(RL)方法将树搜索与深度学习相结合,在搜索非常大的离散动作空间方面取得了显著的成功,例如在国际象棋、将棋和围棋中。然而,许多现实世界的材料发现和设计应用涉及多维搜索问题和具有连续动作空间的学习领域。探索材料的高维势能模型就是一个例子。传统上,这些搜索非常耗时(对于单个体系统通常需要数年时间),并且受到人类直觉和/或专业知识的驱动,最近也受到全局/局部优化搜索的驱动,这些搜索存在收敛问题和/或不能很好地扩展到搜索维度。在这里,我们背离了离散动作和其他基于梯度的方法,引入了一种基于决策树的 RL 策略,该策略结合了经过修改的奖励,以提高探索能力、在播放过程中进行高效采样,以及采用“窗口缩放方案”进行增强利用,从而能够对连续动作空间问题进行高效且可扩展的搜索。我们使用高维人工景观和控制 RL 问题,成功地将我们的方法与流行的全局优化方案和最先进的策略梯度方法进行了基准测试。我们证明了它在参数化潜在模型(基于物理的和高维神经网络)方面的有效性,这些模型适用于周期表中 54 种不同的元素系统以及合金。我们分析了潜在空间中不同元素的误差趋势,并将其归因于元素结构多样性和元素能量表面的平滑度。总的来说,我们的 RL 策略将适用于许多其他涉及连续动作空间搜索的物理科学问题。