• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于非凸优化的具有方差缩减的分布式随机梯度跟踪算法

Distributed Stochastic Gradient Tracking Algorithm With Variance Reduction for Non-Convex Optimization.

作者信息

Jiang Xia, Zeng Xianlin, Sun Jian, Chen Jie

出版信息

IEEE Trans Neural Netw Learn Syst. 2023 Sep;34(9):5310-5321. doi: 10.1109/TNNLS.2022.3170944. Epub 2023 Sep 1.

DOI:10.1109/TNNLS.2022.3170944
PMID:35536804
Abstract

This article proposes a distributed stochastic algorithm with variance reduction for general smooth non-convex finite-sum optimization, which has wide applications in signal processing and machine learning communities. In distributed setting, a large number of samples are allocated to multiple agents in the network. Each agent computes local stochastic gradient and communicates with its neighbors to seek for the global optimum. In this article, we develop a modified variance reduction technique to deal with the variance introduced by stochastic gradients. Combining gradient tracking and variance reduction techniques, this article proposes a distributed stochastic algorithm, gradient tracking algorithm with variance reduction (GT-VR), to solve large-scale non-convex finite-sum optimization over multiagent networks. A complete and rigorous proof shows that the GT-VR algorithm converges to the first-order stationary points with O(1/k) convergence rate. In addition, we provide the complexity analysis of the proposed algorithm. Compared with some existing first-order methods, the proposed algorithm has a lower O(PMϵ) gradient complexity under some mild condition. By comparing state-of-the-art algorithms and GT-VR in numerical simulations, we verify the efficiency of the proposed algorithm.

摘要

本文提出了一种用于一般光滑非凸有限和优化的具有方差缩减的分布式随机算法,该算法在信号处理和机器学习领域有广泛应用。在分布式环境中,大量样本被分配给网络中的多个智能体。每个智能体计算局部随机梯度,并与邻居进行通信以寻求全局最优解。在本文中,我们开发了一种改进的方差缩减技术来处理随机梯度引入的方差。结合梯度跟踪和方差缩减技术,本文提出了一种分布式随机算法——带方差缩减的梯度跟踪算法(GT-VR),用于解决多智能体网络上的大规模非凸有限和优化问题。一个完整且严格的证明表明,GT-VR算法以O(1/k)的收敛速率收敛到一阶驻点。此外,我们还给出了所提算法的复杂度分析。与一些现有的一阶方法相比,在所提算法在一些温和条件下具有更低的O(PMϵ)梯度复杂度。通过在数值模拟中比较现有最优算法和GT-VR,我们验证了所提算法的有效性。

相似文献

1
Distributed Stochastic Gradient Tracking Algorithm With Variance Reduction for Non-Convex Optimization.用于非凸优化的具有方差缩减的分布式随机梯度跟踪算法
IEEE Trans Neural Netw Learn Syst. 2023 Sep;34(9):5310-5321. doi: 10.1109/TNNLS.2022.3170944. Epub 2023 Sep 1.
2
Distributed Stochastic Proximal Algorithm With Random Reshuffling for Nonsmooth Finite-Sum Optimization.用于非光滑有限和优化的带随机重排的分布式随机近端算法
IEEE Trans Neural Netw Learn Syst. 2024 Mar;35(3):4082-4096. doi: 10.1109/TNNLS.2022.3201711. Epub 2024 Feb 29.
3
Stochastic Strongly Convex Optimization via Distributed Epoch Stochastic Gradient Algorithm.通过分布式轮次随机梯度算法实现的随机强凸优化
IEEE Trans Neural Netw Learn Syst. 2021 Jun;32(6):2344-2357. doi: 10.1109/TNNLS.2020.3004723. Epub 2021 Jun 2.
4
A multivariate adaptive gradient algorithm with reduced tuning efforts.一种具有较少调谐工作的多元自适应梯度算法。
Neural Netw. 2022 Aug;152:499-509. doi: 10.1016/j.neunet.2022.05.016. Epub 2022 May 21.
5
Faster First-Order Methods for Stochastic Non-Convex Optimization on Riemannian Manifolds.黎曼流形上随机非凸优化的更快一阶方法
IEEE Trans Pattern Anal Mach Intell. 2021 Feb;43(2):459-472. doi: 10.1109/TPAMI.2019.2933841. Epub 2021 Jan 8.
6
Shuffling-type gradient method with bandwidth-based step sizes for finite-sum optimization.基于带宽的步长的有限和优化的洗牌型梯度方法。
Neural Netw. 2024 Nov;179:106514. doi: 10.1016/j.neunet.2024.106514. Epub 2024 Jul 6.
7
Painless Stochastic Conjugate Gradient for Large-Scale Machine Learning.用于大规模机器学习的无痛随机共轭梯度法
IEEE Trans Neural Netw Learn Syst. 2024 Oct;35(10):14645-14658. doi: 10.1109/TNNLS.2023.3280826. Epub 2024 Oct 7.
8
Improved Variance Reduction Methods for Riemannian Non-Convex Optimization.用于黎曼非凸优化的改进方差缩减方法
IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):7610-7623. doi: 10.1109/TPAMI.2021.3112139. Epub 2022 Oct 4.
9
On Consensus-Optimality Trade-offs in Collaborative Deep Learning.关于协作深度学习中的共识最优性权衡
Front Artif Intell. 2021 Sep 14;4:573731. doi: 10.3389/frai.2021.573731. eCollection 2021.
10
Variance Reduced Methods for Non-Convex Composition Optimization.非凸复合优化的方差缩减方法
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):5813-5825. doi: 10.1109/TPAMI.2021.3071594. Epub 2022 Aug 4.