Li Jyun-Wei, Teng Yu-Chieh, Nimura Shinji, Manie Yibeltal Chanie, Yazdandoost Kamya Yekeh, Tanaka Kazuki, Inohara Ryo, Tsuritani Takehiro, Peng Peng-Chun
Opt Lett. 2024 Feb 1;49(3):666-669. doi: 10.1364/OL.502638.
We successfully demonstrated an intelligent adaptive beam alignment scheme using a reinforcement learning (RL) algorithm integrated with an 8 × 8 photonic array antenna operating in the 40 GHz millimeter wave (MMW) band. In our proposed scheme, the three key elements of RL: state, action, and reward, are represented as the phase values in the photonic array antenna, phase changes with specified steps, and an obtained error vector magnitude (EVM) value, respectively. Furthermore, thanks to the Q-table, the RL agent can effectively choose the most suitable action based on its prior experiences. As a result, the proposed scheme autonomously achieves the best EVM performance by determining the optimal phase. In this Letter, we verify the capability of the proposed scheme in single- and multiple-user scenarios and experimentally demonstrate the performance of beam alignment to the user's location optimized by the RL algorithm. The achieved results always meet the signal quality requirement specified by the 3rd Generation Partnership Project (3GPP) criterion for 64-QAM orthogonal frequency division multiplexing (OFDM).
我们成功展示了一种智能自适应波束对准方案,该方案使用了一种强化学习(RL)算法,并与一个工作在40吉赫兹毫米波(MMW)频段的8×8光子阵列天线相结合。在我们提出的方案中,强化学习的三个关键要素:状态、动作和奖励,分别由光子阵列天线中的相位值、以指定步长的相位变化以及获得的误差向量幅度(EVM)值来表示。此外,借助Q表,强化学习智能体可以根据其先前的经验有效地选择最合适的动作。结果,所提出的方案通过确定最佳相位自主实现了最佳的EVM性能。在本信函中,我们验证了所提出方案在单用户和多用户场景中的能力,并通过实验证明了波束对准到由RL算法优化的用户位置的性能。所取得的结果始终满足第三代合作伙伴计划(3GPP)针对64正交幅度调制(QAM)正交频分复用(OFDM)标准规定的信号质量要求。