National Intellectual Property Administration, Beijing City, China.
PLoS One. 2020 Jul 2;15(7):e0235447. doi: 10.1371/journal.pone.0235447. eCollection 2020.
To explore the application of deep neural networks (DNNs) and deep reinforcement learning (DRL) in wireless communication and accelerate the development of the wireless communication industry.
This study proposes a simple cognitive radio scenario consisting of only one primary user and one secondary user. The secondary user attempts to share spectrum resources with the primary user. An intelligent power algorithm model based on DNNs and DRL is constructed. Then, the MATLAB platform is utilized to simulate the model.
In the performance analysis of the algorithm model under different strategies, it is found that the second power control strategy is more conservative than the first. In the loss function, the second power control strategy has experienced more iterations than the first. In terms of success rate, the second power control strategy has more iterations than the first. In the average number of transmissions, they show the same changing trend, but the success rate can reach 1. In comparison with the traditional distributed clustering and power control (DCPC) algorithm, it is obvious that the convergence rate of the algorithm in this research is higher. The proposed DQN algorithm based on DRL only needs several steps to achieve convergence, which verifies its effectiveness.
By applying DNNs and DRL to model algorithms constructed in wireless scenarios, the success rate is higher and the convergence rate is faster, which can provide experimental basis for the improvement of later wireless communication networks.
探索深度神经网络(DNN)和深度强化学习(DRL)在无线通信中的应用,加速无线通信产业的发展。
本研究提出了一个仅由一个主用户和一个次用户组成的简单认知无线电场景。次用户试图与主用户共享频谱资源。构建了基于 DNN 和 DRL 的智能功率算法模型。然后,利用 MATLAB 平台对模型进行仿真。
在不同策略下的算法模型性能分析中,发现第二功率控制策略比第一功率控制策略更为保守。在损失函数中,第二功率控制策略的迭代次数比第一功率控制策略多。在成功率方面,第二功率控制策略的迭代次数比第一功率控制策略多。在平均传输次数上,它们呈现出相同的变化趋势,但成功率可以达到 1。与传统的分布式聚类和功率控制(DCPC)算法相比,本研究中算法的收敛速度明显更快。基于 DRL 的 DQN 算法仅需几步即可达到收敛,验证了其有效性。
通过将 DNN 和 DRL 应用于无线场景中的模型算法,提高了成功率,加快了收敛速度,可为后续无线通信网络的改进提供实验依据。