• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于强化学习的多射频系统鲁棒资源管理。

Reinforcement-Learning-Based Robust Resource Management for Multi-Radio Systems.

机构信息

Manufacturing, Materials and Mechatronics, School of Engineering, STEM College, RMIT University, 124 La Trobe St., Melbourne, VIC 3000, Australia.

出版信息

Sensors (Basel). 2023 May 17;23(10):4821. doi: 10.3390/s23104821.

DOI:10.3390/s23104821
PMID:37430736
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10223095/
Abstract

The advent of the Internet of Things (IoT) has triggered an increased demand for sensing devices with multiple integrated wireless transceivers. These platforms often support the advantageous use of multiple radio technologies to exploit their differing characteristics. Intelligent radio selection techniques allow these systems to become highly adaptive, ensuring more robust and reliable communications under dynamic channel conditions. In this paper, we focus on the wireless links between devices equipped by deployed operating personnel and intermediary access-point infrastructure. We use multi-radio platforms and wireless devices with multiple and diverse transceiver technologies to produce robust and reliable links through the adaptive control of available transceivers. In this work, the term 'robust' refers to communications that can be maintained despite changes in the environmental and radio conditions, i.e., during periods of interference caused by non-cooperative actors or multi-path or fading conditions in the physical environment. In this paper, a multi-objective reinforcement learning (MORL) framework is applied to address a multi-radio selection and power control problem. We propose independent reward functions to manage the trade-off between the conflicting objectives of minimised power consumption and maximised bit rate. We also adopt an adaptive exploration strategy for learning a robust behaviour policy and compare its online performance to conventional methods. An extension to the multi-objective state-action-reward-state-action (SARSA) algorithm is proposed to implement this adaptive exploration strategy. When applying adaptive exploration to the extended multi-objective SARSA algorithm, we achieve a 20% increase in the F1 score in comparison to one with decayed exploration policies.

摘要

物联网 (IoT) 的出现引发了对具有多个集成无线收发器的传感设备的需求增长。这些平台通常支持多种无线电技术的有利利用,以利用它们不同的特性。智能无线电选择技术允许这些系统具有高度适应性,确保在动态信道条件下实现更稳健和可靠的通信。在本文中,我们专注于配备有部署操作人员和中间接入点基础设施的设备之间的无线链路。我们使用多无线电平台和具有多个不同收发器技术的无线设备,通过对可用收发器的自适应控制来生成稳健可靠的链路。在这项工作中,“稳健”一词是指即使在环境和无线电条件发生变化的情况下,即非合作行为者引起的干扰或物理环境中的多径或衰落条件期间,也能保持通信。在本文中,应用多目标强化学习 (MORL) 框架来解决多无线电选择和功率控制问题。我们提出了独立的奖励函数来管理最小化功耗和最大化比特率这两个冲突目标之间的权衡。我们还采用了自适应探索策略来学习稳健的行为策略,并将其在线性能与传统方法进行比较。提出了一种扩展的多目标状态-动作-奖励-状态-动作 (SARSA) 算法来实现这种自适应探索策略。在将自适应探索应用于扩展的多目标 SARSA 算法时,与使用衰减探索策略相比,我们的 F1 得分提高了 20%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/70b5/10223095/6a9ca0f4093d/sensors-23-04821-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/70b5/10223095/75ecd987db42/sensors-23-04821-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/70b5/10223095/cb59c34bb839/sensors-23-04821-g002a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/70b5/10223095/6a9ca0f4093d/sensors-23-04821-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/70b5/10223095/75ecd987db42/sensors-23-04821-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/70b5/10223095/cb59c34bb839/sensors-23-04821-g002a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/70b5/10223095/6a9ca0f4093d/sensors-23-04821-g003.jpg

相似文献

1
Reinforcement-Learning-Based Robust Resource Management for Multi-Radio Systems.基于强化学习的多射频系统鲁棒资源管理。
Sensors (Basel). 2023 May 17;23(10):4821. doi: 10.3390/s23104821.
2
Joint Beamforming, Power Allocation, and Splitting Control for SWIPT-Enabled IoT Networks with Deep Reinforcement Learning and Game Theory.基于深度强化学习和博弈论的支持同时无线信息与能量传输的物联网网络的联合波束成形、功率分配和分割控制
Sensors (Basel). 2022 Mar 17;22(6):2328. doi: 10.3390/s22062328.
3
A Cascaded Multi-Agent Reinforcement Learning-Based Resource Allocation for Cellular-V2X Vehicular Platooning Networks.基于级联多智能体强化学习的蜂窝车联网车辆编队网络资源分配
Sensors (Basel). 2024 Aug 30;24(17):5658. doi: 10.3390/s24175658.
4
Fuzzy-Assisted Mobile Edge Orchestrator and SARSA Learning for Flexible Offloading in Heterogeneous IoT Environment.模糊辅助移动边缘协调器和 SARSA 学习在异构物联网环境中的灵活卸载。
Sensors (Basel). 2022 Jun 23;22(13):4727. doi: 10.3390/s22134727.
5
Offloading and Transmission Strategies for IoT Edge Devices and Networks.物联网边缘设备和网络的卸载和传输策略。
Sensors (Basel). 2019 Feb 18;19(4):835. doi: 10.3390/s19040835.
6
Intelligent Rapid Adaptive Offloading Algorithm for Computational Services in Dynamic Internet of Things System.动态物联网系统中计算服务的智能快速自适应卸载算法
Sensors (Basel). 2019 Aug 4;19(15):3423. doi: 10.3390/s19153423.
7
Self-Learning Power Control in Wireless Sensor Networks.无线传感器网络中的自学习功率控制
Sensors (Basel). 2018 Jan 27;18(2):375. doi: 10.3390/s18020375.
8
An Effective Spectrum Handoff Based on Reinforcement Learning for Target Channel Selection in the Industrial Internet of Things.基于强化学习的工业物联网目标信道选择的有效频谱切换。
Sensors (Basel). 2019 Mar 21;19(6):1395. doi: 10.3390/s19061395.
9
Dynamic Resource Optimization for Energy-Efficient 6G-IoT Ecosystems.面向高能效 6G-IoT 生态系统的动态资源优化。
Sensors (Basel). 2023 May 12;23(10):4711. doi: 10.3390/s23104711.
10
Multi-Agent Reinforcement Learning for Joint Cooperative Spectrum Sensing and Channel Access in Cognitive UAV Networks.多智能体强化学习在认知无人机网络中的联合协作频谱感知和信道接入。
Sensors (Basel). 2022 Feb 20;22(4):1651. doi: 10.3390/s22041651.

本文引用的文献

1
Global Emergency System Based on WPAN and LPWAN Hybrid Networks.基于 WPAN 和 LPWAN 混合网络的全球应急系统。
Sensors (Basel). 2022 Oct 18;22(20):7921. doi: 10.3390/s22207921.
2
Array programming with NumPy.使用 NumPy 进行数组编程。
Nature. 2020 Sep;585(7825):357-362. doi: 10.1038/s41586-020-2649-2. Epub 2020 Sep 16.
3
Self-Learning Power Control in Wireless Sensor Networks.无线传感器网络中的自学习功率控制
Sensors (Basel). 2018 Jan 27;18(2):375. doi: 10.3390/s18020375.
4
"Notice of Violation of IEEE Publication Principles" Multiobjective Reinforcement Learning: A Comprehensive Overview.《违反IEEE出版原则通知》 多目标强化学习:全面概述
IEEE Trans Cybern. 2013 Apr 29. doi: 10.1109/TSMCC.2013.2249512.
5
Online learning of shaping rewards in reinforcement learning.强化学习中的塑造奖励在线学习。
Neural Netw. 2010 May;23(4):541-50. doi: 10.1016/j.neunet.2010.01.001. Epub 2010 Jan 11.