• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于大规模机器学习的无痛随机共轭梯度法

Painless Stochastic Conjugate Gradient for Large-Scale Machine Learning.

作者信息

Yang Zhuang

出版信息

IEEE Trans Neural Netw Learn Syst. 2024 Oct;35(10):14645-14658. doi: 10.1109/TNNLS.2023.3280826. Epub 2024 Oct 7.

DOI:10.1109/TNNLS.2023.3280826
PMID:37285250
Abstract

Conjugate gradient (CG), as an effective technique to speed up gradient descent algorithms, has shown great potential and has widely been used for large-scale machine-learning problems. However, CG and its variants have not been devised for the stochastic setting, which makes them extremely unstable, and even leads to divergence when using noisy gradients. This article develops a novel class of stable stochastic CG (SCG) algorithms with a faster convergence rate via the variance-reduced technique and an adaptive step size rule in the mini-batch setting. Actually, replacing the use of a line search in the CG-type approaches which is time-consuming, or even fails for SCG, this article considers using the random stabilized Barzilai-Borwein (RSBB) method to obtain an online step size. We rigorously analyze the convergence properties of the proposed algorithms and show that the proposed algorithms attain a linear convergence rate for both the strongly convex and nonconvex settings. Also, we show that the total complexity of the proposed algorithms matches that of modern stochastic optimization algorithms under different cases. Scores of numerical experiments on machine-learning problems demonstrate that the proposed algorithms outperform state-of-the-art stochastic optimization algorithms.

摘要

共轭梯度(CG)作为一种加速梯度下降算法的有效技术,已显示出巨大潜力,并广泛应用于大规模机器学习问题。然而,CG及其变体尚未针对随机设置设计,这使得它们极其不稳定,甚至在使用有噪声的梯度时会导致发散。本文通过在小批量设置中采用方差缩减技术和自适应步长规则,开发了一类收敛速度更快的新型稳定随机共轭梯度(SCG)算法。实际上,由于在CG型方法中使用线搜索既耗时,甚至对SCG而言还会失败,本文考虑使用随机稳定的Barzilai-Borwein(RSBB)方法来获得在线步长。我们严格分析了所提出算法的收敛性质,并表明所提出的算法在强凸和非凸设置下均达到线性收敛速度。此外,我们表明所提出算法的总复杂度在不同情况下与现代随机优化算法相匹配。在机器学习问题上进行的大量数值实验表明,所提出的算法优于现有最先进的随机优化算法。

相似文献

1
Painless Stochastic Conjugate Gradient for Large-Scale Machine Learning.用于大规模机器学习的无痛随机共轭梯度法
IEEE Trans Neural Netw Learn Syst. 2024 Oct;35(10):14645-14658. doi: 10.1109/TNNLS.2023.3280826. Epub 2024 Oct 7.
2
Faster Stochastic Quasi-Newton Methods.更快的随机拟牛顿法
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4388-4397. doi: 10.1109/TNNLS.2021.3056947. Epub 2022 Aug 31.
3
A Minibatch Proximal Stochastic Recursive Gradient Algorithm Using a Trust-Region-Like Scheme and Barzilai-Borwein Stepsizes.一种使用类信赖域方案和巴齐莱-博温步长的小批量近端随机递归梯度算法。
IEEE Trans Neural Netw Learn Syst. 2021 Oct;32(10):4627-4638. doi: 10.1109/TNNLS.2020.3025383. Epub 2021 Oct 5.
4
Stochastic Conjugate Gradient Algorithm With Variance Reduction.具有方差缩减的随机共轭梯度算法
IEEE Trans Neural Netw Learn Syst. 2019 May;30(5):1360-1369. doi: 10.1109/TNNLS.2018.2868835. Epub 2018 Sep 27.
5
Distributed Stochastic Gradient Tracking Algorithm With Variance Reduction for Non-Convex Optimization.用于非凸优化的具有方差缩减的分布式随机梯度跟踪算法
IEEE Trans Neural Netw Learn Syst. 2023 Sep;34(9):5310-5321. doi: 10.1109/TNNLS.2022.3170944. Epub 2023 Sep 1.
6
Stochastic learning via optimizing the variational inequalities.通过优化变分不等式进行随机学习。
IEEE Trans Neural Netw Learn Syst. 2014 Oct;25(10):1769-78. doi: 10.1109/TNNLS.2013.2294741.
7
Learning Rates for Nonconvex Pairwise Learning.非凸对学习的学习率。
IEEE Trans Pattern Anal Mach Intell. 2023 Aug;45(8):9996-10011. doi: 10.1109/TPAMI.2023.3259324. Epub 2023 Jun 30.
8
Stochastic momentum methods for non-convex learning without bounded assumptions.无界假设下非凸学习的随机动量方法。
Neural Netw. 2023 Aug;165:830-845. doi: 10.1016/j.neunet.2023.06.021. Epub 2023 Jun 23.
9
Online Stochastic DCA With Applications to Principal Component Analysis.应用于主成分分析的在线随机判别式坐标下降法
IEEE Trans Neural Netw Learn Syst. 2024 May;35(5):7035-7047. doi: 10.1109/TNNLS.2022.3213558. Epub 2024 May 2.
10
Dualityfree Methods for Stochastic Composition Optimization.随机组合优化的无对偶方法
IEEE Trans Neural Netw Learn Syst. 2019 Apr;30(4):1205-1217. doi: 10.1109/TNNLS.2018.2866699. Epub 2018 Sep 12.

引用本文的文献

1
Probabilistic single-particle cryo-EM ab initio 3D reconstruction in SIMPLE.在SIMPLE中进行概率单粒子冷冻电镜从头三维重建。
Acta Crystallogr D Struct Biol. 2025 Aug 1;81(Pt 8):396-409. doi: 10.1107/S2059798325005686. Epub 2025 Jul 7.