• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

A Hybrid Stochastic-Deterministic Minibatch Proximal Gradient Method for Efficient Optimization and Generalization.

作者信息

Zhou Pan, Yuan Xiaotong, Lin Zhouchen, Hoi Steven

出版信息

IEEE Trans Pattern Anal Mach Intell. 2021 Jun 8;PP. doi: 10.1109/TPAMI.2021.3087328.

DOI:10.1109/TPAMI.2021.3087328
PMID:34101583
Abstract

Despite the success of stochastic variance-reduced gradient (SVRG) algorithms in solving large-scale problems, their stochastic gradient complexity often scales linearly with data size and is expensive for huge data. Accordingly, we propose a hybrid stochastic-deterministic minibatch proximal gradient(HSDMPG) algorithm for strongly convex problems with linear prediction structure, e.g.least squares and logistic/softmax regression. HSDMPGenjoys improved computational complexity that is data-size-independent for large-scale problems. It iteratively samples an evolvingminibatch of individual losses to estimate the original problem, and efficiently minimizes the sampled smaller-sized subproblems. For strongly convex loss of n components, HSDMPGattains an ϵ-optimization-error within [Formula: see text] stochastic gradient evaluations, where κ is condition number, ζ = 1 for quadratic loss and ζ = 2 for generic loss. For large-scale problems, our complexity outperforms those of SVRG-type algorithms with/without dependence on data size. Particularly, when ϵ = O(1/√n) which matches the intrinsic excess error of a learning model and is sufficient for generalization, our complexity for quadratic and generic losses is respectively O (nlog(n)) and O (nlog(n)), which for the first time achieves optimal generalization in less than a single pass over data. Besides, we extend HSDMPGto online strongly convex problems and prove its higher efficiency over the prior algorithms. Numerical results demonstrate the computational advantages of~HSDMPG.

摘要

相似文献

1
A Hybrid Stochastic-Deterministic Minibatch Proximal Gradient Method for Efficient Optimization and Generalization.
IEEE Trans Pattern Anal Mach Intell. 2021 Jun 8;PP. doi: 10.1109/TPAMI.2021.3087328.
2
A Minibatch Proximal Stochastic Recursive Gradient Algorithm Using a Trust-Region-Like Scheme and Barzilai-Borwein Stepsizes.一种使用类信赖域方案和巴齐莱-博温步长的小批量近端随机递归梯度算法。
IEEE Trans Neural Netw Learn Syst. 2021 Oct;32(10):4627-4638. doi: 10.1109/TNNLS.2020.3025383. Epub 2021 Oct 5.
3
Stochastic quasi-gradient methods: variance reduction via Jacobian sketching.随机拟梯度方法:通过雅可比矩阵草图实现方差缩减。
Math Program. 2021;188(1):135-192. doi: 10.1007/s10107-020-01506-0. Epub 2020 May 12.
4
Weighted SGD for ℓ Regression with Randomized Preconditioning.用于带随机预处理的ℓ回归的加权随机梯度下降法。
Proc Annu ACM SIAM Symp Discret Algorithms. 2016 Jan;2016:558-569. doi: 10.1137/1.9781611974331.ch41.
5
Accelerated Variance Reduction Stochastic ADMM for Large-Scale Machine Learning.用于大规模机器学习的加速方差缩减随机交替方向乘子法
IEEE Trans Pattern Anal Mach Intell. 2021 Dec;43(12):4242-4255. doi: 10.1109/TPAMI.2020.3000512. Epub 2021 Nov 3.
6
Faster First-Order Methods for Stochastic Non-Convex Optimization on Riemannian Manifolds.黎曼流形上随机非凸优化的更快一阶方法
IEEE Trans Pattern Anal Mach Intell. 2021 Feb;43(2):459-472. doi: 10.1109/TPAMI.2019.2933841. Epub 2021 Jan 8.
7
Gradient Descent Ascent for Minimax Problems on Riemannian Manifolds.黎曼流形上最小最大化问题的梯度上升法。
IEEE Trans Pattern Anal Mach Intell. 2023 Jul;45(7):8466-8476. doi: 10.1109/TPAMI.2023.3234160. Epub 2023 Jun 5.
8
An accelerated proximal gradient algorithm for singly linearly constrained quadratic programs with box constraints.一种用于具有盒约束的单线性约束二次规划的加速近端梯度算法。
ScientificWorldJournal. 2013 Oct 7;2013:246596. doi: 10.1155/2013/246596. eCollection 2013.
9
Painless Stochastic Conjugate Gradient for Large-Scale Machine Learning.用于大规模机器学习的无痛随机共轭梯度法
IEEE Trans Neural Netw Learn Syst. 2024 Oct;35(10):14645-14658. doi: 10.1109/TNNLS.2023.3280826. Epub 2024 Oct 7.
10
Distributed Stochastic Gradient Tracking Algorithm With Variance Reduction for Non-Convex Optimization.用于非凸优化的具有方差缩减的分布式随机梯度跟踪算法
IEEE Trans Neural Netw Learn Syst. 2023 Sep;34(9):5310-5321. doi: 10.1109/TNNLS.2022.3170944. Epub 2023 Sep 1.