• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

深度强化学习辅助的下行链路正交频分多址协作系统资源分配优化

Deep Reinforcement Learning-Assisted Optimization for Resource Allocation in Downlink OFDMA Cooperative Systems.

作者信息

Tefera Mulugeta Kassaw, Zhang Shengbing, Jin Zengwang

机构信息

School of Cybersecurity, Northwestern Polytechnical University, Xi'an 710072, China.

出版信息

Entropy (Basel). 2023 Feb 24;25(3):413. doi: 10.3390/e25030413.

DOI:10.3390/e25030413
PMID:36981302
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10047118/
Abstract

This paper considers a downlink resource-allocation problem in distributed interference orthogonal frequency-division multiple access (OFDMA) systems under maximal power constraints. As the upcoming fifth-generation (5G) wireless networks are increasingly complex and heterogeneous, it is challenging for resource allocation tasks to optimize the system performance metrics and guarantee user service requests simultaneously. Because of the non-convex optimization problems, using existing approaches to find the optimal resource allocation is computationally expensive. Recently, model-free reinforcement learning (RL) techniques have become alternative approaches in wireless networks to solve non-convex and NP-hard optimization problems. In this paper, we study a deep Q-learning (DQL)-based approach to address the optimization of transmit power control for users in multi-cell interference networks. In particular, we have applied a DQL algorithm for resource allocation to maximize the overall system throughput subject to the maximum power and SINR constraints in a flat frequency channel. We first formulate the optimization problem as a non-cooperative game model, where the multiple BSs compete for spectral efficiencies by improving their achievable utility functions while ensuring the quality of service (QoS) requirements to the corresponding receivers. Then, we develop a DRL-based resource allocation model to maximize the system throughput while satisfying the power and spectral efficiency requirements. In this setting, we define the state-action spaces and the reward function to explore the possible actions and learning outcomes. The numerical simulations demonstrate that the proposed DQL-based scheme outperforms the traditional model-based solution.

摘要

本文研究了在最大功率约束下分布式干扰正交频分多址(OFDMA)系统中的下行链路资源分配问题。随着即将到来的第五代(5G)无线网络日益复杂和异构,资源分配任务要同时优化系统性能指标并保证用户服务请求具有挑战性。由于存在非凸优化问题,使用现有方法寻找最优资源分配在计算上代价高昂。近来,无模型强化学习(RL)技术已成为无线网络中解决非凸和NP难优化问题的替代方法。在本文中,我们研究一种基于深度Q学习(DQL)的方法来解决多小区干扰网络中用户的发射功率控制优化问题。具体而言,我们应用一种DQL算法进行资源分配,以在平坦频率信道中的最大功率和信干噪比(SINR)约束下最大化整体系统吞吐量。我们首先将优化问题表述为一个非合作博弈模型,其中多个基站通过改善其可实现的效用函数来竞争频谱效率,同时确保对相应接收机的服务质量(QoS)要求。然后,我们开发一个基于深度强化学习(DRL)的资源分配模型,以在满足功率和频谱效率要求的同时最大化系统吞吐量。在此设置下,我们定义状态 - 动作空间和奖励函数以探索可能的动作和学习结果。数值模拟表明,所提出的基于DQL的方案优于传统的基于模型的解决方案。

相似文献

1
Deep Reinforcement Learning-Assisted Optimization for Resource Allocation in Downlink OFDMA Cooperative Systems.深度强化学习辅助的下行链路正交频分多址协作系统资源分配优化
Entropy (Basel). 2023 Feb 24;25(3):413. doi: 10.3390/e25030413.
2
Joint Beamforming, Power Allocation, and Splitting Control for SWIPT-Enabled IoT Networks with Deep Reinforcement Learning and Game Theory.基于深度强化学习和博弈论的支持同时无线信息与能量传输的物联网网络的联合波束成形、功率分配和分割控制
Sensors (Basel). 2022 Mar 17;22(6):2328. doi: 10.3390/s22062328.
3
Rate Adaptive Based Resource Allocation with Proportional Fairness Constraints in OFDMA Systems.基于速率自适应的正交频分多址(OFDMA)系统中具有比例公平性约束的资源分配
Sensors (Basel). 2015 Sep 25;15(10):24996-5014. doi: 10.3390/s151024996.
4
Design Optimization of Resource Allocation in OFDMA-Based Cognitive Radio-Enabled Internet of Vehicles (IoVs).基于 OFDMA 的认知无线电增强型车联网 (IoV) 中的资源分配设计优化。
Sensors (Basel). 2020 Nov 9;20(21):6402. doi: 10.3390/s20216402.
5
Joint Optimization of Bandwidth and Power Allocation in Uplink Systems with Deep Reinforcement Learning.基于深度强化学习的上行链路系统中带宽与功率分配的联合优化
Sensors (Basel). 2023 Jul 31;23(15):6822. doi: 10.3390/s23156822.
6
Self-Coexistence among IEEE 802.22 Networks: Distributed Allocation of Power and Channel.IEEE 802.22网络中的自共存:功率与信道的分布式分配
Sensors (Basel). 2017 Dec 7;17(12):2838. doi: 10.3390/s17122838.
7
Mobility-Aware Resource Allocation in IoRT Network for Post-Disaster Communications with Parameterized Reinforcement Learning.用于灾后通信的物联网网络中基于参数强化学习的移动感知资源分配
Sensors (Basel). 2023 Jul 17;23(14):6448. doi: 10.3390/s23146448.
8
Dynamic resource allocation for energy-efficient downlink NOMA systems in 5G networks.5G网络中用于节能下行链路非正交多址接入系统的动态资源分配
Heliyon. 2024 Apr 19;10(9):e29956. doi: 10.1016/j.heliyon.2024.e29956. eCollection 2024 May 15.
9
Resource Allocation for Downlink Full-Duplex Cooperative NOMA-Based Cellular System with Imperfect SI Cancellation and Underlaying D2D Communications.下行链路全双工协作 NOMA 蜂窝系统中具有不完善的 SI 消除和底层 D2D 通信的资源分配。
Sensors (Basel). 2021 Apr 14;21(8):2768. doi: 10.3390/s21082768.
10
Power Allocation and Energy Cooperation for UAV-Enabled MmWave Networks: A Multi-Agent Deep Reinforcement Learning Approach.无人机增强毫米波网络的功率分配和能量合作:一种多智能体深度强化学习方法。
Sensors (Basel). 2021 Dec 30;22(1):270. doi: 10.3390/s22010270.

本文引用的文献

1
A Review of Fundamental Optimization Approaches and the Role of AI Enabling Technologies in Physical Layer Security.一种基本优化方法综述及人工智能使能技术在物理层安全中的作用。
Sensors (Basel). 2022 May 9;22(9):3589. doi: 10.3390/s22093589.
2
Deep Reinforcement Learning for Resource Management on Network Slicing: A Survey.深度强化学习在网络切片资源管理中的应用研究综述。
Sensors (Basel). 2022 Apr 15;22(8):3031. doi: 10.3390/s22083031.
3
NPENAS: Neural Predictor Guided Evolution for Neural Architecture Search.NPENAS:用于神经架构搜索的神经预测器引导进化
IEEE Trans Neural Netw Learn Syst. 2023 Nov;34(11):8441-8455. doi: 10.1109/TNNLS.2022.3151160. Epub 2023 Oct 27.
4
Human-level control through deep reinforcement learning.通过深度强化学习实现人类水平的控制。
Nature. 2015 Feb 26;518(7540):529-33. doi: 10.1038/nature14236.