• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

神经进化与梯度下降的对应关系。

Correspondence between neuroevolution and gradient descent.

机构信息

Molecular Foundry, Lawrence Berkeley National Laboratory, 1 Cyclotron Road, Berkeley, CA, 94720, USA.

Department of Physics, University of Ottawa, Ottawa, ON, K1N 6N5, Canada.

出版信息

Nat Commun. 2021 Nov 2;12(1):6317. doi: 10.1038/s41467-021-26568-2.

DOI:10.1038/s41467-021-26568-2
PMID:34728632
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8563972/
Abstract

We show analytically that training a neural network by conditioned stochastic mutation or neuroevolution of its weights is equivalent, in the limit of small mutations, to gradient descent on the loss function in the presence of Gaussian white noise. Averaged over independent realizations of the learning process, neuroevolution is equivalent to gradient descent on the loss function. We use numerical simulation to show that this correspondence can be observed for finite mutations, for shallow and deep neural networks. Our results provide a connection between two families of neural-network training methods that are usually considered to be fundamentally different.

摘要

我们从理论上证明,在小突变的极限下,通过条件随机突变或神经进化对神经网络的权重进行训练,相当于在存在高斯白噪声的情况下对损失函数进行梯度下降。在学习过程的独立实现中进行平均,神经进化相当于对损失函数进行梯度下降。我们通过数值模拟表明,这种对应关系可以在有限的突变下,对于浅层和深层神经网络,都可以观察到。我们的结果为两种通常被认为在根本上不同的神经网络训练方法之间提供了联系。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/4674ede6127b/41467_2021_26568_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/b5a5fa592c82/41467_2021_26568_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/990e124b9c0d/41467_2021_26568_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/20b207748e97/41467_2021_26568_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/e2afd41aae43/41467_2021_26568_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/6a872db8ea3d/41467_2021_26568_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/4674ede6127b/41467_2021_26568_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/b5a5fa592c82/41467_2021_26568_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/990e124b9c0d/41467_2021_26568_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/20b207748e97/41467_2021_26568_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/e2afd41aae43/41467_2021_26568_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/6a872db8ea3d/41467_2021_26568_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f4ac/8563972/4674ede6127b/41467_2021_26568_Fig6_HTML.jpg

相似文献

1
Correspondence between neuroevolution and gradient descent.神经进化与梯度下降的对应关系。
Nat Commun. 2021 Nov 2;12(1):6317. doi: 10.1038/s41467-021-26568-2.
2
Improving the performance of mutation-based evolving artificial neural networks with self-adaptive mutations.利用自适应突变提高基于突变的进化人工神经网络的性能。
PLoS One. 2024 Jul 15;19(7):e0307084. doi: 10.1371/journal.pone.0307084. eCollection 2024.
3
Neuroevolution of a Modular Memory-Augmented Neural Network for Deep Memory Problems.模块化记忆增强神经网络的神经进化用于深度记忆问题。
Evol Comput. 2019 Winter;27(4):639-664. doi: 10.1162/evco_a_00239. Epub 2018 Nov 8.
4
Universality of gradient descent neural network training.梯度下降神经网络训练的通用性。
Neural Netw. 2022 Jun;150:259-273. doi: 10.1016/j.neunet.2022.02.016. Epub 2022 Mar 2.
5
Accelerating deep neural network training with inconsistent stochastic gradient descent.使用不一致随机梯度下降加速深度神经网络训练
Neural Netw. 2017 Sep;93:219-229. doi: 10.1016/j.neunet.2017.06.003. Epub 2017 Jun 16.
6
Is Learning in Biological Neural Networks Based on Stochastic Gradient Descent? An Analysis Using Stochastic Processes.基于随机梯度下降的生物神经网络学习?随机过程分析。
Neural Comput. 2024 Jun 7;36(7):1424-1432. doi: 10.1162/neco_a_01668.
7
Decentralized stochastic sharpness-aware minimization algorithm.去中心化随机锐化感知最小化算法。
Neural Netw. 2024 Aug;176:106325. doi: 10.1016/j.neunet.2024.106325. Epub 2024 Apr 17.
8
Stability analysis of stochastic gradient descent for homogeneous neural networks and linear classifiers.随机梯度下降在同质神经网络和线性分类器中的稳定性分析。
Neural Netw. 2023 Jul;164:382-394. doi: 10.1016/j.neunet.2023.04.028. Epub 2023 Apr 25.
9
Optimizing neural networks for medical data sets: A case study on neonatal apnea prediction.优化神经网络在医学数据集上的应用:以新生儿呼吸暂停预测为例的研究
Artif Intell Med. 2019 Jul;98:59-76. doi: 10.1016/j.artmed.2019.07.008. Epub 2019 Jul 25.
10
Block-cyclic stochastic coordinate descent for deep neural networks.分块循环随机坐标下降法在深度神经网络中的应用。
Neural Netw. 2021 Jul;139:348-357. doi: 10.1016/j.neunet.2021.04.001. Epub 2021 Apr 19.

引用本文的文献

1
Programmable photonic unitary circuits for light computing.用于光计算的可编程光子酉电路。
Nanophotonics. 2025 Feb 14;14(10):1429-1449. doi: 10.1515/nanoph-2024-0602. eCollection 2025 May.
2
Learning protocols for the fast and efficient control of active matter.用于快速高效控制活性物质的学习协议。
Nat Commun. 2024 Oct 23;15(1):9128. doi: 10.1038/s41467-024-52878-2.
3
Stochastic Gradient Descent-like relaxation is equivalent to Metropolis dynamics in discrete optimization and inference problems.在离散优化和推理问题中,类似随机梯度下降的松弛方法等同于 metropolis 动力学。

本文引用的文献

1
Learning to grow: Control of material self-assembly using evolutionary reinforcement learning.学习成长:利用进化强化学习控制材料自组装
Phys Rev E. 2020 May;101(5-1):052604. doi: 10.1103/PhysRevE.101.052604.
2
How to simulate patchy particles.如何模拟斑状颗粒。
Eur Phys J E Soft Matter. 2018 May 14;41(5):59. doi: 10.1140/epje/i2018-11667-x.
3
Deep learning in neural networks: an overview.神经网络中的深度学习:综述。
Sci Rep. 2024 May 21;14(1):11638. doi: 10.1038/s41598-024-62625-8.
4
Learning deep neural networks' architectures using differential evolution. Case study: Medical imaging processing.使用差分进化学习深度神经网络架构。案例研究:医学图像处理。
Comput Biol Med. 2022 Jul;146:105623. doi: 10.1016/j.compbiomed.2022.105623. Epub 2022 May 17.
Neural Netw. 2015 Jan;61:85-117. doi: 10.1016/j.neunet.2014.09.003. Epub 2014 Oct 13.
4
Coarse-grained Monte Carlo simulations of non-equilibrium systems.非平衡系统的粗粒度蒙特卡罗模拟。
J Chem Phys. 2013 Jun 28;138(24):244111. doi: 10.1063/1.4811656.
5
Dynamic Monte Carlo versus Brownian dynamics: A comparison for self-diffusion and crystallization in colloidal fluids.动态蒙特卡洛法与布朗动力学:胶体流体中自扩散和结晶的比较
J Chem Phys. 2010 May 21;132(19):194102. doi: 10.1063/1.3414827.
6
Avoiding unphysical kinetic traps in Monte Carlo simulations of strongly attractive particles.在强吸引粒子的蒙特卡罗模拟中避免非物理动力学陷阱。
J Chem Phys. 2007 Oct 21;127(15):154101. doi: 10.1063/1.2790421.
7
Reversible self-assembly of patchy particles into monodisperse icosahedral clusters.补丁粒子可逆自组装成单分散二十面体簇。
J Chem Phys. 2007 Aug 28;127(8):085106. doi: 10.1063/1.2759922.
8
Revisiting the slow dynamics of a silica melt using Monte Carlo simulations.利用蒙特卡罗模拟重新审视二氧化硅熔体的慢动力学。
Phys Rev E Stat Nonlin Soft Matter Phys. 2007 Jul;76(1 Pt 1):011507. doi: 10.1103/PhysRevE.76.011507. Epub 2007 Jul 16.
9
Rejection-free geometric cluster algorithm for complex fluids.用于复杂流体的无排斥几何聚类算法
Phys Rev Lett. 2004 Jan 23;92(3):035504. doi: 10.1103/PhysRevLett.92.035504.
10
Evolving neural networks through augmenting topologies.通过扩展拓扑结构来演化神经网络。
Evol Comput. 2002 Summer;10(2):99-127. doi: 10.1162/106365602320169811.