• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

具有博弈反馈的分布式在线随机约束凸优化

Distributed Online Stochastic-Constrained Convex Optimization With Bandit Feedback.

作者信息

Wang Cong, Xu Shengyuan, Yuan Deming

出版信息

IEEE Trans Cybern. 2024 Jan;54(1):63-75. doi: 10.1109/TCYB.2022.3177644. Epub 2023 Dec 20.

DOI:10.1109/TCYB.2022.3177644
PMID:35724296
Abstract

This article studies the distributed online stochastic convex optimization problem with the time-varying constraint over a multiagent system constructed by various agents. The sequences of cost functions and constraint functions, both of which have dynamic parameters following time-varying distributions, are unacquainted to the agent ahead of time. Agents in the network are able to interact with their neighbors through a sequence of strongly connected and time-varying graphs. We develop the adaptive distributed bandit primal-dual algorithm whose step size and regularization sequences are adaptive and have no prior knowledge about the total iteration span T . The adaptive distributed bandit primal-dual algorithm applies bandit feedback with a one-point or two-point gradient estimator to evaluate gradient values. It is illustrated in this article that if the drift of the benchmark sequence is sublinear, then the adaptive distributed bandit primal-dual algorithm exhibits sublinear expected dynamic regret and constraint violation using both two kinds of gradient estimator to compute gradient information. We present a numerical experiment to show the performance of the proposed method.

摘要

本文研究了由多个智能体构建的多智能体系统上具有时变约束的分布式在线随机凸优化问题。代价函数序列和约束函数序列都具有遵循时变分布的动态参数,智能体事先并不知晓这些序列。网络中的智能体能够通过一系列强连通且时变的图与邻居进行交互。我们开发了自适应分布式带反馈原始对偶算法,其步长和正则化序列是自适应的,并且对总迭代次数T没有先验知识。自适应分布式带反馈原始对偶算法应用带反馈的单点或两点梯度估计器来评估梯度值。本文表明,如果基准序列的漂移是次线性的,那么使用这两种梯度估计器来计算梯度信息时,自适应分布式带反馈原始对偶算法会表现出次线性的期望动态遗憾和约束违反。我们给出了一个数值实验来展示所提方法的性能。

相似文献

1
Distributed Online Stochastic-Constrained Convex Optimization With Bandit Feedback.具有博弈反馈的分布式在线随机约束凸优化
IEEE Trans Cybern. 2024 Jan;54(1):63-75. doi: 10.1109/TCYB.2022.3177644. Epub 2023 Dec 20.
2
Distributed Online Constrained Optimization With Feedback Delays.具有反馈延迟的分布式在线约束优化
IEEE Trans Neural Netw Learn Syst. 2024 Feb;35(2):1708-1720. doi: 10.1109/TNNLS.2022.3184957. Epub 2024 Feb 5.
3
Online Learning Algorithm for Distributed Convex Optimization With Time-Varying Coupled Constraints and Bandit Feedback.具有时变耦合约束和博弈反馈的分布式凸优化在线学习算法
IEEE Trans Cybern. 2022 Feb;52(2):1009-1020. doi: 10.1109/TCYB.2020.2990796. Epub 2022 Feb 16.
4
Push-Sum Distributed Online Optimization With Bandit Feedback.具有博弈反馈的推和分布式在线优化
IEEE Trans Cybern. 2022 Apr;52(4):2263-2273. doi: 10.1109/TCYB.2020.2999309. Epub 2022 Apr 5.
5
An Adaptive Primal-Dual Subgradient Algorithm for Online Distributed Constrained Optimization.一种用于在线分布式约束优化的自适应对偶子梯度算法。
IEEE Trans Cybern. 2018 Nov;48(11):3045-3055. doi: 10.1109/TCYB.2017.2755720. Epub 2017 Oct 5.
6
Distributed Online Learning Algorithm for Noncooperative Games Over Unbalanced Digraphs.非平衡有向图上非合作博弈的分布式在线学习算法
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):15846-15856. doi: 10.1109/TNNLS.2023.3290049. Epub 2024 Oct 29.
7
Stochastic Strongly Convex Optimization via Distributed Epoch Stochastic Gradient Algorithm.通过分布式轮次随机梯度算法实现的随机强凸优化
IEEE Trans Neural Netw Learn Syst. 2021 Jun;32(6):2344-2357. doi: 10.1109/TNNLS.2020.3004723. Epub 2021 Jun 2.
8
Distributed Optimization for Graph Matching.用于图匹配的分布式优化
IEEE Trans Cybern. 2023 Aug;53(8):4815-4828. doi: 10.1109/TCYB.2022.3140338. Epub 2023 Jul 18.
9
Distributed Frank-Wolfe Solver for Stochastic Optimization With Coupled Inequality Constraints.用于具有耦合不等式约束的随机优化的分布式Frank-Wolfe求解器
IEEE Trans Neural Netw Learn Syst. 2025 May;36(5):7858-7872. doi: 10.1109/TNNLS.2024.3423376. Epub 2025 May 2.
10
Event-Triggered Distributed Stochastic Mirror Descent for Convex Optimization.用于凸优化的事件触发分布式随机镜像下降法
IEEE Trans Neural Netw Learn Syst. 2023 Sep;34(9):6480-6491. doi: 10.1109/TNNLS.2021.3137010. Epub 2023 Sep 1.