• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

面向协作自组织网络的分布式机会频谱资源接入模型与算法

Decentralized Opportunistic Spectrum Resources Access Model and Algorithm toward Cooperative Ad-Hoc Networks.

作者信息

Liu Ming, Xu Yang, Mohammed Abdul-Wahid

机构信息

School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan, P.R. China.

出版信息

PLoS One. 2016 Jan 4;11(1):e0145526. doi: 10.1371/journal.pone.0145526. eCollection 2016.

DOI:10.1371/journal.pone.0145526
PMID:26727504
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC4699651/
Abstract

Limited communication resources have gradually become a critical factor toward efficiency of decentralized large scale multi-agent coordination when both system scales up and tasks become more complex. In current researches, due to the agent's limited communication and observational capability, an agent in a decentralized setting can only choose a part of channels to access, but cannot perceive or share global information. Each agent's cooperative decision is based on the partial observation of the system state, and as such, uncertainty in the communication network is unavoidable. In this situation, it is a major challenge working out cooperative decision-making under uncertainty with only a partial observation of the environment. In this paper, we propose a decentralized approach that allows agents cooperatively search and independently choose channels. The key to our design is to build an up-to-date observation for each agent's view so that a local decision model is achievable in a large scale team coordination. We simplify the Dec-POMDP model problem, and each agent can jointly work out its communication policy in order to improve its local decision utilities for the choice of communication resources. Finally, we discuss an implicate resource competition game, and show that, there exists an approximate resources access tradeoff balance between agents. Based on this discovery, the tradeoff between real-time decision-making and the efficiency of cooperation using these channels can be well improved.

摘要

当系统规模扩大且任务变得更加复杂时,有限的通信资源已逐渐成为影响分散式大规模多智能体协调效率的关键因素。在当前的研究中,由于智能体的通信和观测能力有限,处于分散环境中的智能体只能选择部分信道进行访问,而无法感知或共享全局信息。每个智能体的协作决策都是基于对系统状态的部分观测,因此,通信网络中的不确定性是不可避免的。在这种情况下,仅通过对环境的部分观测来解决不确定性下的协作决策是一项重大挑战。在本文中,我们提出了一种分散式方法,允许智能体进行协作搜索并独立选择信道。我们设计的关键在于为每个智能体的视角构建最新的观测,以便在大规模团队协调中实现局部决策模型。我们简化了分散式部分可观测马尔可夫决策过程(Dec-POMDP)模型问题,每个智能体可以共同制定其通信策略,以提高其在选择通信资源时的局部决策效用。最后,我们讨论了一个隐含的资源竞争博弈,并表明,智能体之间存在近似的资源访问权衡平衡。基于这一发现,可以很好地改善实时决策与使用这些信道的合作效率之间的权衡。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/6268dc9f1ceb/pone.0145526.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/6048fb3319c2/pone.0145526.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/35397ac99fce/pone.0145526.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/d76599546ebf/pone.0145526.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/864e3e9531c5/pone.0145526.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/970b206cdebf/pone.0145526.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/854bd2d6b0ea/pone.0145526.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/6268dc9f1ceb/pone.0145526.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/6048fb3319c2/pone.0145526.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/35397ac99fce/pone.0145526.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/d76599546ebf/pone.0145526.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/864e3e9531c5/pone.0145526.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/970b206cdebf/pone.0145526.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/854bd2d6b0ea/pone.0145526.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/45c4/4699651/6268dc9f1ceb/pone.0145526.g007.jpg

相似文献

1
Decentralized Opportunistic Spectrum Resources Access Model and Algorithm toward Cooperative Ad-Hoc Networks.面向协作自组织网络的分布式机会频谱资源接入模型与算法
PLoS One. 2016 Jan 4;11(1):e0145526. doi: 10.1371/journal.pone.0145526. eCollection 2016.
2
Modeling and Planning with Macro-Actions in Decentralized POMDPs.分散式部分可观察马尔可夫决策过程中宏动作的建模与规划
J Artif Intell Res. 2019;64:817-859. doi: 10.1613/jair.1.11418. Epub 2019 Mar 25.
3
Decentralized Policy Coordination in Mobile Sensing with Consensual Communication.基于共识通信的移动感知中的分散式政策协调。
Sensors (Basel). 2022 Dec 7;22(24):9584. doi: 10.3390/s22249584.
4
Coordination as inference in multi-agent reinforcement learning.多智能体强化学习中的协调作为推理。
Neural Netw. 2024 Apr;172:106101. doi: 10.1016/j.neunet.2024.106101. Epub 2024 Jan 11.
5
Game-theoretic approach for improving cooperation in wireless multihop networks.用于改善无线多跳网络中协作的博弈论方法。
IEEE Trans Syst Man Cybern B Cybern. 2010 Jun;40(3):559-74. doi: 10.1109/TSMCB.2010.2042593. Epub 2010 Mar 4.
6
Learning Multi-Agent Cooperation via Considering Actions of Teammates.通过考虑队友的行动来学习多智能体合作。
IEEE Trans Neural Netw Learn Syst. 2024 Aug;35(8):11553-11564. doi: 10.1109/TNNLS.2023.3262921. Epub 2024 Aug 5.
7
Forward and Backward Bellman Equations Improve the Efficiency of the EM Algorithm for DEC-POMDP.前向和后向贝尔曼方程提高了用于去中心化部分可观测马尔可夫决策过程的期望最大化算法的效率。
Entropy (Basel). 2021 Apr 29;23(5):551. doi: 10.3390/e23050551.
8
Measuring team situation awareness in decentralized command and control environments.在分散式指挥与控制环境中测量团队态势感知。
Ergonomics. 2006;49(12-13):1312-25. doi: 10.1080/00140130600612788.
9
Decentralized Bayesian search using approximate dynamic programming methods.使用近似动态规划方法的分散式贝叶斯搜索
IEEE Trans Syst Man Cybern B Cybern. 2008 Aug;38(4):970-5. doi: 10.1109/TSMCB.2008.928180.
10
Decentralized robust adaptive control for the multiagent system consensus problem using neural networks.基于神经网络的多智能体系统一致性问题的分布式鲁棒自适应控制
IEEE Trans Syst Man Cybern B Cybern. 2009 Jun;39(3):636-47. doi: 10.1109/TSMCB.2008.2007810. Epub 2009 Jan 23.

引用本文的文献

1
Sample-efficient multi-agent reinforcement learning with masked reconstruction.基于掩蔽重建的高效多智能体强化学习。
PLoS One. 2023 Sep 14;18(9):e0291545. doi: 10.1371/journal.pone.0291545. eCollection 2023.
2
Multi-agent reinforcement learning with approximate model learning for competitive games.多智能体强化学习与近似模型学习在竞争性游戏中的应用。
PLoS One. 2019 Sep 11;14(9):e0222215. doi: 10.1371/journal.pone.0222215. eCollection 2019.
3
Continuous patrolling in uncertain environment with the UAV swarm.在不确定环境下使用无人机群进行持续巡逻。

本文引用的文献

1
Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning.通过分布式循环Q学习实现多机器人软管运输与部署学习
PLoS One. 2015 Jul 9;10(7):e0127129. doi: 10.1371/journal.pone.0127129. eCollection 2015.
2
A comprehensive review of swarm optimization algorithms.群体优化算法的全面综述。
PLoS One. 2015 May 18;10(5):e0122827. doi: 10.1371/journal.pone.0122827. eCollection 2015.
3
Evolutionary game theory: Temporal and spatial effects beyond replicator dynamics.进化博弈论:复制者动态之外的时空效应。
PLoS One. 2018 Aug 24;13(8):e0202328. doi: 10.1371/journal.pone.0202328. eCollection 2018.
4
Semantic Agent-Based Service Middleware and Simulation for Smart Cities.基于语义智能体的智慧城市服务中间件与仿真
Sensors (Basel). 2016 Dec 21;16(12):2200. doi: 10.3390/s16122200.
5
A Matrix-Based Proactive Data Relay Algorithm for Large Distributed Sensor Networks.一种用于大型分布式传感器网络的基于矩阵的主动数据中继算法。
Sensors (Basel). 2016 Aug 16;16(8):1300. doi: 10.3390/s16081300.
Phys Life Rev. 2009 Dec;6(4):208-49. doi: 10.1016/j.plrev.2009.08.001. Epub 2009 Aug 7.