• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过双重协作约束协调多智能体强化学习

Coordinating Multi-Agent Reinforcement Learning via Dual Collaborative Constraints.

作者信息

Li Chao, Dong Shaokang, Yang Shangdong, Hu Yujing, Li Wenbin, Gao Yang

机构信息

State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing, 210023, China.

School of Computer Science, Nanjing University of Posts and Telecommunications, Nanjing, 210023, China.

出版信息

Neural Netw. 2025 Feb;182:106858. doi: 10.1016/j.neunet.2024.106858. Epub 2024 Nov 12.

DOI:10.1016/j.neunet.2024.106858
PMID:39550797
Abstract

Many real-world multi-agent tasks exhibit a nearly decomposable structure, where interactions among agents within the same interaction set are strong while interactions between different sets are relatively weak. Efficiently modeling the nearly decomposable structure and leveraging it to coordinate agents can enhance the learning efficiency of multi-agent reinforcement learning algorithms for cooperative tasks, while existing works typically fail. To overcome this limitation, this paper proposes a novel algorithm named Dual Collaborative Constraints (DCC) that identifies the interaction sets as subtasks and achieves both intra-subtask and inter-subtask coordination. Specifically, DCC employs a bi-level structure to periodically distribute agents into multiple subtasks, and proposes both local and global collaborative constraints based on mutual information to facilitate both intra-subtask and inter-subtask coordination among agents. These two constraints ensure that agents within the same subtask reach a consensus on their local action selections and all of them select superior joint actions that maximize the overall task performance. Experimentally, we evaluate DCC on various cooperative multi-agent tasks, and its superior performance against multiple state-of-the-art baselines demonstrates its effectiveness.

摘要

许多现实世界中的多智能体任务呈现出近乎可分解的结构,即同一交互集内的智能体之间的交互很强,而不同集之间的交互相对较弱。有效地对近乎可分解的结构进行建模并利用它来协调智能体,可以提高用于合作任务的多智能体强化学习算法的学习效率,而现有工作通常无法做到这一点。为了克服这一限制,本文提出了一种名为双重协作约束(DCC)的新算法,该算法将交互集识别为子任务,并实现子任务内和子任务间的协调。具体来说,DCC采用双层结构将智能体定期分配到多个子任务中,并基于互信息提出局部和全局协作约束,以促进智能体之间的子任务内和子任务间协调。这两个约束确保同一子任务内的智能体在其局部动作选择上达成共识,并且它们都选择能使整体任务性能最大化的最优联合动作。通过实验,我们在各种合作多智能体任务上评估了DCC,其相对于多个最先进基线的优越性能证明了其有效性。

相似文献

1
Coordinating Multi-Agent Reinforcement Learning via Dual Collaborative Constraints.通过双重协作约束协调多智能体强化学习
Neural Netw. 2025 Feb;182:106858. doi: 10.1016/j.neunet.2024.106858. Epub 2024 Nov 12.
2
Constraining an Unconstrained Multi-agent Policy with offline data.使用离线数据约束无约束多智能体策略。
Neural Netw. 2025 Jun;186:107253. doi: 10.1016/j.neunet.2025.107253. Epub 2025 Feb 13.
3
Multi-Task Multi-Agent Reinforcement Learning With Interaction and Task Representations.具有交互和任务表示的多任务多智能体强化学习
IEEE Trans Neural Netw Learn Syst. 2025 Jul;36(7):13431-13445. doi: 10.1109/TNNLS.2024.3475216.
4
LJIR: Learning Joint-Action Intrinsic Reward in cooperative multi-agent reinforcement learning.LJIR:在合作多智能体强化学习中学习联合行动内在奖励
Neural Netw. 2023 Oct;167:450-459. doi: 10.1016/j.neunet.2023.08.016. Epub 2023 Aug 22.
5
MuDE: Multi-agent decomposed reward-based exploration.MuDE:基于多代理分解奖励的探索。
Neural Netw. 2024 Nov;179:106565. doi: 10.1016/j.neunet.2024.106565. Epub 2024 Jul 22.
6
TIMAR: Transition-informed representation for sample-efficient multi-agent reinforcement learning.TIMAR:用于样本高效多智能体强化学习的转换感知表示
Neural Netw. 2025 Apr;184:107081. doi: 10.1016/j.neunet.2024.107081. Epub 2024 Dec 31.
7
GeneWorker: An end-to-end robotic reinforcement learning approach with collaborative generator and worker networks.基因工作者:一种具有协作生成器和工作网络的端到端机器人强化学习方法。
Neural Netw. 2024 Oct;178:106472. doi: 10.1016/j.neunet.2024.106472. Epub 2024 Jun 18.
8
Optimistic sequential multi-agent reinforcement learning with motivational communication.带有激励性沟通的乐观序贯多智能体强化学习。
Neural Netw. 2024 Nov;179:106547. doi: 10.1016/j.neunet.2024.106547. Epub 2024 Jul 22.
9
Collaborative hunting in artificial agents with deep reinforcement learning.深度强化学习中的人工代理协同捕猎。
Elife. 2024 May 7;13:e85694. doi: 10.7554/eLife.85694.
10
Generative subgoal oriented multi-agent reinforcement learning through potential field.基于势场的面向生成子目标的多智能体强化学习。
Neural Netw. 2024 Nov;179:106552. doi: 10.1016/j.neunet.2024.106552. Epub 2024 Jul 17.