• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于物联网的强化学习使用概率模型通过计算智能确定广泛探索,以用于下一代技术。

IoT-Based Reinforcement Learning Using Probabilistic Model for Determining Extensive Exploration through Computational Intelligence for Next-Generation Techniques.

机构信息

Birla Global University, Gothapatna, Bhubaneswar, Odisha, India.

School of Computing Science & Engineering, Department of CSE, Galgotias University, Greater Noida, UP, India.

出版信息

Comput Intell Neurosci. 2023 Oct 10;2023:5113417. doi: 10.1155/2023/5113417. eCollection 2023.

DOI:10.1155/2023/5113417
PMID:37854640
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10581845/
Abstract

Computing intelligence is built on several learning and optimization techniques. Incorporating cutting-edge learning techniques to balance the interaction between exploitation and exploration is therefore an inspiring field, especially when it is combined with IoT. The reinforcement learning techniques created in recent years have largely focused on incorporating deep learning technology to improve the generalization skills of the algorithm while ignoring the issue of detecting and taking full advantage of the dilemma. To increase the effectiveness of exploration, a deep reinforcement algorithm based on computational intelligence is proposed in this study, using intelligent sensors and the Bayesian approach. In addition, the technique for computing the posterior distribution of parameters in Bayesian linear regression is expanded to nonlinear models such as artificial neural networks. The Bayesian Bootstrap Deep -Network (BBDQN) algorithm is created by combining the bootstrapped DQN with the recommended computing technique. Finally, tests in two scenarios demonstrate that, when faced with severe exploration problems, BBDQN outperforms DQN and bootstrapped DQN in terms of exploration efficiency.

摘要

计算智能建立在几种学习和优化技术之上。因此,将最先进的学习技术融入到平衡开发与探索之间的交互中是一个令人振奋的领域,尤其是当它与物联网结合时。近年来创建的强化学习技术主要侧重于结合深度学习技术来提高算法的泛化能力,而忽略了检测和充分利用困境的问题。为了提高探索的效果,本研究提出了一种基于计算智能的深度强化算法,该算法使用智能传感器和贝叶斯方法。此外,还将贝叶斯线性回归中参数后验分布的计算技术扩展到人工神经网络等非线性模型。通过将带自举的 DQN 与推荐的计算技术相结合,创建了贝叶斯自举深度网络(BBDQN)算法。最后,在两个场景中的测试表明,在面临严重的探索问题时,BBDQN 在探索效率方面优于 DQN 和带自举的 DQN。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/6a701ba5b29b/CIN2023-5113417.alg.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/087eeaa530cd/CIN2023-5113417.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/2031fe88e6d1/CIN2023-5113417.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/806a3bbf38c2/CIN2023-5113417.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/e81103456b81/CIN2023-5113417.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/5664b3452733/CIN2023-5113417.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/6a701ba5b29b/CIN2023-5113417.alg.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/087eeaa530cd/CIN2023-5113417.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/2031fe88e6d1/CIN2023-5113417.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/806a3bbf38c2/CIN2023-5113417.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/e81103456b81/CIN2023-5113417.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/5664b3452733/CIN2023-5113417.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1d58/10581845/6a701ba5b29b/CIN2023-5113417.alg.001.jpg

相似文献

1
IoT-Based Reinforcement Learning Using Probabilistic Model for Determining Extensive Exploration through Computational Intelligence for Next-Generation Techniques.基于物联网的强化学习使用概率模型通过计算智能确定广泛探索,以用于下一代技术。
Comput Intell Neurosci. 2023 Oct 10;2023:5113417. doi: 10.1155/2023/5113417. eCollection 2023.
2
Deep-Reinforcement-Learning-Based IoT Sensor Data Cleaning Framework for Enhanced Data Analytics.基于深度强化学习的物联网传感器数据清理框架,用于增强数据分析。
Sensors (Basel). 2023 Feb 5;23(4):1791. doi: 10.3390/s23041791.
3
A comprehensive review and analysis of supervised-learning and soft computing techniques for stress diagnosis in humans.对人类应激诊断的监督学习和软计算技术的全面回顾和分析。
Comput Biol Med. 2021 Jul;134:104450. doi: 10.1016/j.compbiomed.2021.104450. Epub 2021 May 5.
4
Dynamic Bayesian network structure learning based on an improved bacterial foraging optimization algorithm.基于改进细菌觅食优化算法的动态贝叶斯网络结构学习。
Sci Rep. 2024 Apr 9;14(1):8266. doi: 10.1038/s41598-024-58806-0.
5
Intelligent Task Dispatching and Scheduling Using a Deep Q-Network in a Cluster Edge Computing System.在集群边缘计算系统中使用深度Q网络的智能任务调度与分配
Sensors (Basel). 2022 May 28;22(11):4098. doi: 10.3390/s22114098.
6
Advances in Computational Intelligence Techniques-Based Multi-Intersection Querying Theory for Efficient QoS in the Next Generation Internet of Things.基于计算智能技术的多交叉查询理论在下一代物联网中的高效服务质量研究进展。
Comput Intell Neurosci. 2023 Jul 8;2023:1388425. doi: 10.1155/2023/1388425. eCollection 2023.
7
A Study of Two-Way Short- and Long-Term Memory Network Intelligent Computing IoT Model-Assisted Home Education Attention Mechanism.双向短时和长时记忆网络智能计算物联网模型辅助家庭教育注意力机制的研究。
Comput Intell Neurosci. 2021 Dec 21;2021:3587884. doi: 10.1155/2021/3587884. eCollection 2021.
8
Deep Reinforcement Learning-Empowered Resource Allocation for Mobile Edge Computing in Cellular V2X Networks.深度强化学习助力蜂窝车联网网络中移动边缘计算的资源分配
Sensors (Basel). 2021 Jan 7;21(2):372. doi: 10.3390/s21020372.
9
Event-Driven Deep Learning for Edge Intelligence (EDL-EI).面向边缘智能的事件驱动深度学习(EDL-EI)。
Sensors (Basel). 2021 Sep 8;21(18):6023. doi: 10.3390/s21186023.
10
Multisource Transfer Double DQN Based on Actor Learning.基于演员学习的多源转移双 DQN。
IEEE Trans Neural Netw Learn Syst. 2018 Jun;29(6):2227-2238. doi: 10.1109/TNNLS.2018.2806087.

本文引用的文献

1
SASDL and RBATQ: Sparse Autoencoder With Swarm Based Deep Learning and Reinforcement Based Q-Learning for EEG Classification.SASDL和RBATQ:基于群体深度学习和基于强化Q学习的稀疏自动编码器用于脑电图分类
IEEE Open J Eng Med Biol. 2022 Mar 23;3:58-68. doi: 10.1109/OJEMB.2022.3161837. eCollection 2022.
2
Weak Human Preference Supervision for Deep Reinforcement Learning.深度强化学习中的弱人类偏好监督
IEEE Trans Neural Netw Learn Syst. 2021 Dec;32(12):5369-5378. doi: 10.1109/TNNLS.2021.3084198. Epub 2021 Nov 30.
3
USV Formation and Path-Following Control via Deep Reinforcement Learning With Random Braking.
基于随机制动的深度强化学习的无人水面航行器编队形成与路径跟踪控制
IEEE Trans Neural Netw Learn Syst. 2021 Dec;32(12):5468-5478. doi: 10.1109/TNNLS.2021.3068762. Epub 2021 Nov 30.
4
Feature Control as Intrinsic Motivation for Hierarchical Reinforcement Learning.特征控制作为分层强化学习的内在动机。
IEEE Trans Neural Netw Learn Syst. 2019 Nov;30(11):3409-3418. doi: 10.1109/TNNLS.2019.2891792. Epub 2019 Jan 29.