Geng Xuan, Zheng Yahong Rosa
IEEE Trans Neural Netw Learn Syst. 2023 Dec;34(12):10626-10637. doi: 10.1109/TNNLS.2022.3170050. Epub 2023 Nov 30.
This article proposes a novel deep-reinforcement learning-based medium access control (DL-MAC) protocol for underwater acoustic networks (UANs) where one agent node employing the proposed DL-MAC protocol coexists with other nodes employing traditional protocols, such as time division multiple access (TDMA) or q -Aloha. The DL-MAC agent learns to exploit the large propagation delays inherent in underwater acoustic communications to improve system throughput by either a synchronous or an asynchronous transmission mode. In the sync-DL-MAC protocol, the agent action space is transmission or no transmission, while in the async-DL-MAC, the agent can also vary the start time in each transmission time slot to further exploit the spatiotemporal uncertainty of the UANs. The deep Q -learning algorithm is applied to both sync-DL-MAC and async-DL-MAC agents to learn the optimal policies. A theoretical analysis and computer simulations demonstrate the performance gain obtained by both DL-MAC protocols. The async-DL-MAC protocol outperforms the sync-DL-MAC protocol significantly in sum throughput and packet success rate by adjusting the transmission start time and reducing the length of time slot.
本文针对水下声学网络(UAN)提出了一种基于深度强化学习的新型介质访问控制(DL-MAC)协议,其中采用所提DL-MAC协议的一个代理节点与采用传统协议(如时分多址(TDMA)或q -Aloha)的其他节点共存。DL-MAC代理通过同步或异步传输模式,学习利用水下声学通信固有的大传播延迟来提高系统吞吐量。在同步DL-MAC协议中,代理动作空间为传输或不传输,而在异步DL-MAC协议中,代理还可在每个传输时隙中改变开始时间,以进一步利用UAN的时空不确定性。深度Q学习算法应用于同步DL-MAC和异步DL-MAC代理,以学习最优策略。理论分析和计算机仿真证明了两种DL-MAC协议所获得的性能增益。异步DL-MAC协议通过调整传输开始时间和减少时隙长度,在总吞吐量和分组成功率方面显著优于同步DL-MAC协议。