School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin 541004, China.
School of Computer Science and Engineering, Northeastern University, Shenyang 110819, China.
Sensors (Basel). 2022 Oct 24;22(21):8139. doi: 10.3390/s22218139.
Optical transport networks (OTNs) are widely used in backbone- and metro-area transmission networks to increase network transmission capacity. In the OTN, it is particularly crucial to rationally allocate routes and maximize network capacities. By employing deep reinforcement learning (DRL)- and software-defined networking (SDN)-based solutions, the capacity of optical networks can be effectively increased. However, because most DRL-based routing optimization methods have low sample usage and difficulty in coping with sudden network connectivity changes, converging in software-defined OTN scenarios is challenging. Additionally, the generalization ability of these methods is weak. This paper proposes an ensembles- and message-passing neural-network-based Deep Q-Network (EMDQN) method for optical network routing optimization to address this problem. To effectively explore the environment and improve agent performance, the multiple EMDQN agents select actions based on the highest upper-confidence bounds. Furthermore, the EMDQN agent captures the network's spatial feature information using a message passing neural network (MPNN)-based DRL policy network, which enables the DRL agent to have generalization capability. The experimental results show that the EMDQN algorithm proposed in this paper performs better in terms of convergence. EMDQN effectively improves the throughput rate and link utilization of optical networks and has better generalization capabilities.
光传送网(OTN)广泛应用于骨干网和城域网传输网络,以提高网络传输容量。在 OTN 中,合理分配路由和最大化网络容量尤为重要。通过采用基于深度强化学习(DRL)和软件定义网络(SDN)的解决方案,可以有效提高光网络的容量。然而,由于大多数基于 DRL 的路由优化方法样本利用率低,难以应对突发的网络连接变化,因此在软件定义的 OTN 场景中收敛是具有挑战性的。此外,这些方法的泛化能力较弱。针对这一问题,本文提出了一种基于集合和消息传递神经网络的深度 Q 网络(EMDQN)方法,用于光网络路由优化。为了有效地探索环境并提高代理性能,多个 EMDQN 代理根据最高置信上限选择动作。此外,EMDQN 代理使用基于消息传递神经网络(MPNN)的 DRL 策略网络捕获网络的空间特征信息,从而使 DRL 代理具有泛化能力。实验结果表明,本文提出的 EMDQN 算法在收敛性方面表现更好。EMDQN 有效地提高了光网络的吞吐量和链路利用率,具有更好的泛化能力。