• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于速度范围的奖励塑造技术,用于通过激光雷达传感器和深度强化学习实现有效的无地图导航。

Velocity range-based reward shaping technique for effective map-less navigation with LiDAR sensor and deep reinforcement learning.

作者信息

Lee HyeokSoo, Jeong Jongpil

机构信息

Department of Smart Factory Convergence, AI Factory Lab, Sungkyunkwan University, Suwon, Republic of Korea.

Research & Development Team, THiRA-UTECH Co., Ltd., Seoul, Republic of Korea.

出版信息

Front Neurorobot. 2023 Sep 6;17:1210442. doi: 10.3389/fnbot.2023.1210442. eCollection 2023.

DOI:10.3389/fnbot.2023.1210442
PMID:37744086
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10512054/
Abstract

In recent years, sensor components similar to human sensory functions have been rapidly developed in the hardware field, enabling the acquisition of information at a level beyond that of humans, and in the software field, artificial intelligence technology has been utilized to enable cognitive abilities and decision-making such as prediction, analysis, and judgment. These changes are being utilized in various industries and fields. In particular, new hardware and software technologies are being rapidly applied to robotics products, showing a level of performance and completeness that was previously unimaginable. In this paper, we researched the topic of establishing an optimal path plan for autonomous driving using LiDAR sensors and deep reinforcement learning in a workplace without map and grid coordinates for mobile robots, which are widely used in logistics and manufacturing sites. For this purpose, we reviewed the hardware configuration of mobile robots capable of autonomous driving, checked the characteristics of the main core sensors, and investigated the core technologies of autonomous driving. In addition, we reviewed the appropriate deep reinforcement learning algorithm to realize the autonomous driving of mobile robots, defined a deep neural network for autonomous driving data conversion, and defined a reward function for path planning. The contents investigated in this paper were built into a simulation environment to verify the autonomous path planning through experiment, and an additional reward technique "Velocity Range-based Evaluation Method" was proposed for further improvement of performance indicators required in the real field, and the effectiveness was verified. The simulation environment and detailed results of experiments are described in this paper, and it is expected as guidance and reference research for applying these technologies in the field.

摘要

近年来,在硬件领域,类似于人类感官功能的传感器组件得到了迅速发展,使得信息采集能力超越了人类水平;在软件领域,人工智能技术被用于实现诸如预测、分析和判断等认知能力与决策。这些变化正被应用于各个行业和领域。特别是,新的硬件和软件技术正在迅速应用于机器人产品,展现出前所未有的性能和完整性水平。在本文中,我们研究了在物流和制造场所广泛使用的、无地图和网格坐标的工作场所中,利用激光雷达传感器和深度强化学习为移动机器人建立自动驾驶最优路径规划的课题。为此,我们回顾了具备自动驾驶能力的移动机器人的硬件配置,检查了主要核心传感器的特性,并研究了自动驾驶的核心技术。此外,我们回顾了实现移动机器人自动驾驶的合适深度强化学习算法,定义了用于自动驾驶数据转换的深度神经网络,并定义了路径规划的奖励函数。本文所研究的内容被构建到一个模拟环境中,通过实验验证自主路径规划,并提出了一种额外的奖励技术“基于速度范围的评估方法”,以进一步提高实际场景中所需性能指标,验证了其有效性。本文描述了模拟环境和详细的实验结果,期望能为这些技术在该领域的应用提供指导和参考研究。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/1206400b287b/fnbot-17-1210442-g0014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/0aded0b72ecf/fnbot-17-1210442-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/210fbaafe575/fnbot-17-1210442-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/4b967a8d35b8/fnbot-17-1210442-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/09bf8c307c9b/fnbot-17-1210442-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/07b60ffd2771/fnbot-17-1210442-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/5b1e09ae5ca8/fnbot-17-1210442-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/b39815551c5d/fnbot-17-1210442-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/135dc1d1fedb/fnbot-17-1210442-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/de0f8438aa62/fnbot-17-1210442-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/f0a8c82723d5/fnbot-17-1210442-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/78d262a63751/fnbot-17-1210442-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/6e1f2f41f13a/fnbot-17-1210442-g0012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/ddb1f8de0d3c/fnbot-17-1210442-g0013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/1206400b287b/fnbot-17-1210442-g0014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/0aded0b72ecf/fnbot-17-1210442-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/210fbaafe575/fnbot-17-1210442-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/4b967a8d35b8/fnbot-17-1210442-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/09bf8c307c9b/fnbot-17-1210442-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/07b60ffd2771/fnbot-17-1210442-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/5b1e09ae5ca8/fnbot-17-1210442-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/b39815551c5d/fnbot-17-1210442-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/135dc1d1fedb/fnbot-17-1210442-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/de0f8438aa62/fnbot-17-1210442-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/f0a8c82723d5/fnbot-17-1210442-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/78d262a63751/fnbot-17-1210442-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/6e1f2f41f13a/fnbot-17-1210442-g0012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/ddb1f8de0d3c/fnbot-17-1210442-g0013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dcb4/10512054/1206400b287b/fnbot-17-1210442-g0014.jpg

相似文献

1
Velocity range-based reward shaping technique for effective map-less navigation with LiDAR sensor and deep reinforcement learning.基于速度范围的奖励塑造技术,用于通过激光雷达传感器和深度强化学习实现有效的无地图导航。
Front Neurorobot. 2023 Sep 6;17:1210442. doi: 10.3389/fnbot.2023.1210442. eCollection 2023.
2
Autonomous Navigation by Mobile Robot with Sensor Fusion Based on Deep Reinforcement Learning.基于深度强化学习的传感器融合移动机器人自主导航
Sensors (Basel). 2024 Jun 16;24(12):3895. doi: 10.3390/s24123895.
3
Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward Environments.基于深度确定性策略梯度的稀疏奖励环境下移动机器人自主驾驶。
Sensors (Basel). 2022 Dec 7;22(24):9574. doi: 10.3390/s22249574.
4
The Impact of LiDAR Configuration on Goal-Based Navigation within a Deep Reinforcement Learning Framework.激光雷达配置对深度强化学习框架内基于目标的导航的影响
Sensors (Basel). 2023 Dec 9;23(24):9732. doi: 10.3390/s23249732.
5
Autonomous Driving of Mobile Robots in Dynamic Environments Based on Deep Deterministic Policy Gradient: Reward Shaping and Hindsight Experience Replay.基于深度确定性策略梯度的动态环境中移动机器人自主驾驶:奖励塑造与事后经验回放
Biomimetics (Basel). 2024 Jan 13;9(1):0. doi: 10.3390/biomimetics9010051.
6
Autonomous Navigation System of Greenhouse Mobile Robot Based on 3D Lidar and 2D Lidar SLAM.基于3D激光雷达和2D激光雷达同步定位与地图构建的温室移动机器人自主导航系统
Front Plant Sci. 2022 Mar 10;13:815218. doi: 10.3389/fpls.2022.815218. eCollection 2022.
7
Research and Implementation of Autonomous Navigation for Mobile Robots Based on SLAM Algorithm under ROS.基于ROS下SLAM算法的移动机器人自主导航研究与实现
Sensors (Basel). 2022 May 31;22(11):4172. doi: 10.3390/s22114172.
8
Research on obstacle avoidance optimization and path planning of autonomous vehicles based on attention mechanism combined with multimodal information decision-making thoughts of robots.基于注意力机制结合机器人多模态信息决策思想的自动驾驶车辆避障优化与路径规划研究
Front Neurorobot. 2023 Sep 22;17:1269447. doi: 10.3389/fnbot.2023.1269447. eCollection 2023.
9
Deep reinforcement learning-aided autonomous navigation with landmark generators.基于地标生成器的深度强化学习辅助自主导航。
Front Neurorobot. 2023 Aug 22;17:1200214. doi: 10.3389/fnbot.2023.1200214. eCollection 2023.
10
The Path Planning of Mobile Robot by Neural Networks and Hierarchical Reinforcement Learning.基于神经网络和分层强化学习的移动机器人路径规划
Front Neurorobot. 2020 Oct 2;14:63. doi: 10.3389/fnbot.2020.00063. eCollection 2020.

本文引用的文献

1
The Path Planning of Mobile Robot by Neural Networks and Hierarchical Reinforcement Learning.基于神经网络和分层强化学习的移动机器人路径规划
Front Neurorobot. 2020 Oct 2;14:63. doi: 10.3389/fnbot.2020.00063. eCollection 2020.