• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

Fully Decentralized Multiagent Communication via Causal Inference.

作者信息

Wang Han, Yu Yang, Jiang Yuan

出版信息

IEEE Trans Neural Netw Learn Syst. 2023 Dec;34(12):10193-10202. doi: 10.1109/TNNLS.2022.3165114. Epub 2023 Nov 30.

DOI:10.1109/TNNLS.2022.3165114
PMID:35452391
Abstract

Many real-world tasks can be cast into multiagent (MA) reinforcement learning problems, and most algorithms in this field obey to the centralized learning and decentralized execution framework. However, enforcing centralized learning is impractical in many scenarios. Because it requires integrating the information from agents, while agents may not hope to share local information due to the issue of privacy. Thus, this article proposes a novel approach to achieve fully decentralized learning based on communication among multiple agents via reinforcement learning. Benefiting from causality analysis, an agent will choose the counterfactual that has the most significant influence on communication information of others. We find that this method can be applied in classic or complex MA scenarios and in federated learning domains, which are now attracting much attention.

摘要

相似文献

1
Fully Decentralized Multiagent Communication via Causal Inference.
IEEE Trans Neural Netw Learn Syst. 2023 Dec;34(12):10193-10202. doi: 10.1109/TNNLS.2022.3165114. Epub 2023 Nov 30.
2
Multiagent Continual Coordination via Progressive Task Contextualization.通过渐进式任务情境化实现多智能体持续协调
IEEE Trans Neural Netw Learn Syst. 2025 Apr;36(4):6326-6340. doi: 10.1109/TNNLS.2024.3394513. Epub 2025 Apr 4.
3
Coordination as inference in multi-agent reinforcement learning.多智能体强化学习中的协调作为推理。
Neural Netw. 2024 Apr;172:106101. doi: 10.1016/j.neunet.2024.106101. Epub 2024 Jan 11.
4
Towards coordinated and robust real-time control: a decentralized approach for combined sewer overflow and urban flooding reduction based on multi-agent reinforcement learning.面向协调稳健的实时控制:基于多智能体强化学习的合流制污水溢流与城市内涝协同减排的分散式方法。
Water Res. 2023 Feb 1;229:119498. doi: 10.1016/j.watres.2022.119498. Epub 2022 Dec 15.
5
Egoism, utilitarianism and egalitarianism in multi-agent reinforcement learning.多智能体强化学习中的利己主义、功利主义和平等主义。
Neural Netw. 2024 Oct;178:106544. doi: 10.1016/j.neunet.2024.106544. Epub 2024 Jul 24.
6
SMIX(λ): Enhancing Centralized Value Functions for Cooperative Multiagent Reinforcement Learning.SMIX(λ):增强用于协作多智能体强化学习的集中式价值函数
IEEE Trans Neural Netw Learn Syst. 2023 Jan;34(1):52-63. doi: 10.1109/TNNLS.2021.3089493. Epub 2023 Jan 5.
7
A Local-and-Global Attention Reinforcement Learning Algorithm for Multiagent Cooperative Navigation.一种用于多智能体协作导航的局部与全局注意力强化学习算法
IEEE Trans Neural Netw Learn Syst. 2024 Jun;35(6):7767-7777. doi: 10.1109/TNNLS.2022.3220798. Epub 2024 Jun 3.
8
Communication-efficient federated learning via knowledge distillation.基于知识蒸馏的高效通信联邦学习。
Nat Commun. 2022 Apr 19;13(1):2032. doi: 10.1038/s41467-022-29763-x.
9
Decentralized federated learning through proxy model sharing.通过代理模型共享的去中心化联邦学习。
Nat Commun. 2023 May 22;14(1):2899. doi: 10.1038/s41467-023-38569-4.
10
HyperComm: Hypergraph-based communication in multi-agent reinforcement learning.超通讯:多智能体强化学习中的基于超图的通讯。
Neural Netw. 2024 Oct;178:106432. doi: 10.1016/j.neunet.2024.106432. Epub 2024 Jun 10.