• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

反向传播算法的动态学习率优化

Dynamic learning rate optimization of the backpropagation algorithm.

作者信息

Yu X H, Chen G A, Cheng S X

机构信息

Dept. of Radio Eng., Southeast Univ., Nanjing.

出版信息

IEEE Trans Neural Netw. 1995;6(3):669-77. doi: 10.1109/72.377972.

DOI:10.1109/72.377972
PMID:18263352
Abstract

It has been observed by many authors that the backpropagation (BP) error surfaces usually consist of a large amount of flat regions as well as extremely steep regions. As such, the BP algorithm with a fixed learning rate will have low efficiency. This paper considers dynamic learning rate optimization of the BP algorithm using derivative information. An efficient method of deriving the first and second derivatives of the objective function with respect to the learning rate is explored, which does not involve explicit calculation of second-order derivatives in weight space, but rather uses the information gathered from the forward and backward propagation, Several learning rate optimization approaches are subsequently established based on linear expansion of the actual outputs and line searches with acceptable descent value and Newton-like methods, respectively. Simultaneous determination of the optimal learning rate and momentum is also introduced by showing the equivalence between the momentum version BP and the conjugate gradient method. Since these approaches are constructed by simple manipulations of the obtained derivatives, the computational and storage burden scale with the network size exactly like the standard BP algorithm, and the convergence of the BP algorithm is accelerated with in a remarkable reduction (typically by factor 10 to 50, depending upon network architectures and applications) in the running time for the overall learning process. Numerous computer simulation results are provided to support the present approaches.

摘要

许多作者已经观察到,反向传播(BP)误差曲面通常由大量的平坦区域以及极其陡峭的区域组成。因此,固定学习率的BP算法效率会很低。本文考虑使用导数信息对BP算法进行动态学习率优化。探索了一种有效推导目标函数关于学习率的一阶和二阶导数的方法,该方法不涉及在权重空间中显式计算二阶导数,而是利用前向和反向传播收集的信息。随后分别基于实际输出的线性展开、具有可接受下降值的线搜索和类牛顿方法建立了几种学习率优化方法。通过展示动量版BP与共轭梯度法之间的等价性,还引入了同时确定最优学习率和动量的方法。由于这些方法是通过对所得导数进行简单操作构建的,计算和存储负担与网络大小的比例关系与标准BP算法完全相同,并且BP算法的收敛速度在整体学习过程的运行时间上有显著降低(通常降低10到50倍,具体取决于网络架构和应用)。提供了大量计算机模拟结果来支持当前方法。

相似文献

1
Dynamic learning rate optimization of the backpropagation algorithm.反向传播算法的动态学习率优化
IEEE Trans Neural Netw. 1995;6(3):669-77. doi: 10.1109/72.377972.
2
Stability analysis of a three-term backpropagation algorithm.一种三项反向传播算法的稳定性分析
Neural Netw. 2005 Dec;18(10):1341-7. doi: 10.1016/j.neunet.2005.04.007. Epub 2005 Aug 30.
3
An accelerated learning algorithm for multilayer perceptrons: optimization layer by layer.一种用于多层感知器的加速学习算法:逐层优化
IEEE Trans Neural Netw. 1995;6(1):31-42. doi: 10.1109/72.363452.
4
Magnified gradient function with deterministic weight modification in adaptive learning.自适应学习中具有确定性权重修改的放大梯度函数
IEEE Trans Neural Netw. 2004 Nov;15(6):1411-23. doi: 10.1109/TNN.2004.836237.
5
The layer-wise method and the backpropagation hybrid approach to learning a feedforward neural network.用于学习前馈神经网络的逐层方法与反向传播混合方法。
IEEE Trans Neural Netw. 2000;11(2):295-305. doi: 10.1109/72.839001.
6
Global convergence of online BP training with dynamic learning rate.在线 BP 训练的动态学习率全局收敛。
IEEE Trans Neural Netw Learn Syst. 2012 Feb;23(2):330-41. doi: 10.1109/TNNLS.2011.2178315.
7
On-line learning algorithms for locally recurrent neural networks.局部递归神经网络的在线学习算法
IEEE Trans Neural Netw. 1999;10(2):253-71. doi: 10.1109/72.750549.
8
A generalized learning paradigm exploiting the structure of feedforward neural networks.一种利用前馈神经网络结构的通用学习范式。
IEEE Trans Neural Netw. 1996;7(6):1450-60. doi: 10.1109/72.548172.
9
A Circuit-Based Neural Network with Hybrid Learning of Backpropagation and Random Weight Change Algorithms.一种基于电路的神经网络,结合了反向传播和随机权重变化算法的混合学习。
Sensors (Basel). 2016 Dec 23;17(1):16. doi: 10.3390/s17010016.
10
Steepest descent with momentum for quadratic functions is a version of the conjugate gradient method.二次函数的带动量的最速下降法是共轭梯度法的一种形式。
Neural Netw. 2004 Jan;17(1):65-71. doi: 10.1016/S0893-6080(03)00170-9.

引用本文的文献

1
Groundwater chemical characteristic analysis and water source identification model study in Gubei coal mine, Northern Anhui Province, China.中国安徽省北部古北煤矿的地下水化学特征分析与水源识别模型研究
Heliyon. 2024 Feb 27;10(5):e26925. doi: 10.1016/j.heliyon.2024.e26925. eCollection 2024 Mar 15.
2
Deep Transfer Learning for Land Use and Land Cover Classification: A Comparative Study.深度学习在土地利用和土地覆盖分类中的应用:一项比较研究。
Sensors (Basel). 2021 Dec 3;21(23):8083. doi: 10.3390/s21238083.
3
A self-adaptive deep learning method for automated eye laterality detection based on color fundus photography.
基于彩色眼底照相的自动眼偏侧性检测的自适应深度学习方法。
PLoS One. 2019 Sep 19;14(9):e0222025. doi: 10.1371/journal.pone.0222025. eCollection 2019.