• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于D2D网络中信道选择和功率控制的联合深度强化学习与无监督学习

Joint Deep Reinforcement Learning and Unsupervised Learning for Channel Selection and Power Control in D2D Networks.

作者信息

Sun Ming, Jin Yanhui, Wang Shumei, Mei Erzhuang

机构信息

College of Computer and Control Engineering, Qiqihar University, Qiqihar 161006, China.

School of Computer and Information Engineering, Harbin University of Commerce, Harbin 150028, China.

出版信息

Entropy (Basel). 2022 Nov 24;24(12):1722. doi: 10.3390/e24121722.

DOI:10.3390/e24121722
PMID:36554127
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9777944/
Abstract

Device-to-device (D2D) technology enables direct communication between devices, which can effectively solve the problem of insufficient spectrum resources in 5G communication technology. Since the channels are shared among multiple D2D user pairs, it may lead to serious interference between D2D user pairs. In order to reduce interference, effectively increase network capacity, and improve wireless spectrum utilization, this paper proposed a distributed resource allocation algorithm with the joint of a deep Q network (DQN) and an unsupervised learning network. Firstly, a DQN algorithm was constructed to solve the channel allocation in the dynamic and unknown environment in a distributed manner. Then, a deep power control neural network with the unsupervised learning strategy was constructed to output an optimized channel power control scheme to maximize the spectrum transmit sum-rate through the corresponding constraint processing. As opposed to traditional centralized approaches that require the collection of instantaneous global network information, the algorithm proposed in this paper used each transmitter as a learning agent to make channel selection and power control through a small amount of state information collected locally. The simulation results showed that the proposed algorithm was more effective in increasing the convergence speed and maximizing the transmit sum-rate than other traditional centralized and distributed algorithms.

摘要

设备到设备(D2D)技术实现了设备之间的直接通信,这可以有效解决5G通信技术中频谱资源不足的问题。由于信道在多个D2D用户对之间共享,这可能导致D2D用户对之间产生严重干扰。为了减少干扰、有效增加网络容量并提高无线频谱利用率,本文提出了一种结合深度Q网络(DQN)和无监督学习网络的分布式资源分配算法。首先,构建了一种DQN算法,以分布式方式解决动态和未知环境中的信道分配问题。然后,构建了一种具有无监督学习策略的深度功率控制神经网络,通过相应的约束处理输出优化的信道功率控制方案,以最大化频谱传输总速率。与需要收集瞬时全局网络信息的传统集中式方法不同,本文提出的算法将每个发射机作为学习代理,通过本地收集的少量状态信息进行信道选择和功率控制。仿真结果表明,与其他传统的集中式和分布式算法相比,该算法在提高收敛速度和最大化传输总速率方面更有效。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/e289500a74d5/entropy-24-01722-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/f226b2fd576d/entropy-24-01722-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/4266b572d438/entropy-24-01722-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/d139974a1bec/entropy-24-01722-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/a3a0fb91bfea/entropy-24-01722-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/024544dab69c/entropy-24-01722-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/ece3a81d12e8/entropy-24-01722-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/8d2125274ca9/entropy-24-01722-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/070a26a0aea1/entropy-24-01722-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/0cd592428c1c/entropy-24-01722-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/bd902cb6ee45/entropy-24-01722-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/4f0fe7c547f9/entropy-24-01722-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/b6fa8bf234c5/entropy-24-01722-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/d5f96a2156b3/entropy-24-01722-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/b61363f3dd09/entropy-24-01722-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/a0570c372344/entropy-24-01722-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/2c1c6d8ac497/entropy-24-01722-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/58d87cb43503/entropy-24-01722-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/e289500a74d5/entropy-24-01722-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/f226b2fd576d/entropy-24-01722-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/4266b572d438/entropy-24-01722-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/d139974a1bec/entropy-24-01722-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/a3a0fb91bfea/entropy-24-01722-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/024544dab69c/entropy-24-01722-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/ece3a81d12e8/entropy-24-01722-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/8d2125274ca9/entropy-24-01722-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/070a26a0aea1/entropy-24-01722-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/0cd592428c1c/entropy-24-01722-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/bd902cb6ee45/entropy-24-01722-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/4f0fe7c547f9/entropy-24-01722-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/b6fa8bf234c5/entropy-24-01722-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/d5f96a2156b3/entropy-24-01722-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/b61363f3dd09/entropy-24-01722-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/a0570c372344/entropy-24-01722-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/2c1c6d8ac497/entropy-24-01722-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/58d87cb43503/entropy-24-01722-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/268e/9777944/e289500a74d5/entropy-24-01722-g018.jpg

相似文献

1
Joint Deep Reinforcement Learning and Unsupervised Learning for Channel Selection and Power Control in D2D Networks.用于D2D网络中信道选择和功率控制的联合深度强化学习与无监督学习
Entropy (Basel). 2022 Nov 24;24(12):1722. doi: 10.3390/e24121722.
2
Two Tier Slicing Resource Allocation Algorithm Based on Deep Reinforcement Learning and Joint Bidding in Wireless Access Networks.基于深度强化学习和联合竞价的无线接入网络双层切片资源分配算法
Sensors (Basel). 2022 May 4;22(9):3495. doi: 10.3390/s22093495.
3
Deep Reinforcement Learning Based Resource Allocation for D2D Communications Underlay Cellular Networks.基于深度强化学习的蜂窝网络中 D2D 通信的资源分配。
Sensors (Basel). 2022 Dec 3;22(23):9459. doi: 10.3390/s22239459.
4
Joint Particle Swarm Optimization of Power and Phase Shift for IRS-Aided D2D Underlaying Cellular Systems.基于 IRS 的 D2D 覆盖蜂窝系统中的功率和相移联合粒子群优化。
Sensors (Basel). 2023 Jun 1;23(11):5266. doi: 10.3390/s23115266.
5
D2D-Assisted Multi-User Cooperative Partial Offloading in MEC Based on Deep Reinforcement Learning.基于深度强化学习的MEC中D2D辅助多用户协作部分卸载
Sensors (Basel). 2022 Sep 15;22(18):7004. doi: 10.3390/s22187004.
6
A Self-Regulating Power-Control Scheme Using Reinforcement Learning for D2D Communication Networks.一种用于D2D通信网络的基于强化学习的自调节功率控制方案。
Sensors (Basel). 2022 Jun 29;22(13):4894. doi: 10.3390/s22134894.
7
A Two-Layer, Energy-Efficient Approach for Joint Power Control and Uplink-Downlink Channel Allocation in D2D Communication.一种用于D2D通信中联合功率控制和上行-下行链路信道分配的两层节能方法。
Sensors (Basel). 2020 Jun 9;20(11):3285. doi: 10.3390/s20113285.
8
Cellular Network Power Allocation Algorithm Based on Deep Reinforcement Learning and Artificial Intelligence.基于深度学习和人工智能的蜂窝网络功率分配算法。
Comput Intell Neurosci. 2022 Jun 24;2022:9456611. doi: 10.1155/2022/9456611. eCollection 2022.
9
Energy-Efficient Resource Allocation Based on Deep Q-Network in V2V Communications.基于深度 Q 网络的车对车通信中的节能资源分配。
Sensors (Basel). 2023 Jan 23;23(3):1295. doi: 10.3390/s23031295.
10
Joint Power Control and Resource Allocation for Optimizing the D2D User Performance of Full-Duplex D2D Underlying Cellular Networks.用于优化全双工D2D底层蜂窝网络中D2D用户性能的联合功率控制与资源分配
Sensors (Basel). 2023 Dec 1;23(23):9549. doi: 10.3390/s23239549.

引用本文的文献

1
Channel and Power Allocation for Multi-Cell NOMA Using Multi-Agent Deep Reinforcement Learning and Unsupervised Learning.基于多智能体深度强化学习和无监督学习的多小区非正交多址接入的信道与功率分配
Sensors (Basel). 2025 Apr 25;25(9):2733. doi: 10.3390/s25092733.