• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

四足机器人的分层强化学习:受限环境中的高效物体操纵

Hierarchical Reinforcement Learning for Quadrupedal Robots: Efficient Object Manipulation in Constrained Environments.

作者信息

Azimi David, Hoseinnezhad Reza

机构信息

School of Information Technology, Deakin University, Victoria 3125, Australia.

School of Engineering, RMIT University, Victoria 3082, Australia.

出版信息

Sensors (Basel). 2025 Mar 4;25(5):1565. doi: 10.3390/s25051565.

DOI:10.3390/s25051565
PMID:40096360
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11902496/
Abstract

This study introduces a hierarchical reinforcement learning (RL) framework tailored to object manipulation tasks by quadrupedal robots, emphasizing their real-world deployment. The proposed approach adopts a sensor-driven control structure capable of addressing challenges in dense and cluttered environments filled with walls and obstacles. A novel reward function is central to the method, incorporating sensor-based obstacle observations to optimize the decision-making. This design minimizes the computational demands while maintaining adaptability and robust functionality. Simulated trials conducted in NVIDIA Isaac Sim, utilizing ANYbotics quadrupedal robots, demonstrated a high manipulation accuracy, with a mean positioning error of 11 cm across object-target distances of up to 10 m. Furthermore, the RL framework effectively integrates path planning in complex environments, achieving energy-efficient and stable operations. These findings establish the framework as a promising approach for advanced robotics requiring versatility, efficiency, and practical deployability.

摘要

本研究介绍了一种专门为四足机器人的物体操纵任务量身定制的分层强化学习(RL)框架,强调其在现实世界中的部署。所提出的方法采用了一种传感器驱动的控制结构,能够应对充满墙壁和障碍物的密集和杂乱环境中的挑战。一种新颖的奖励函数是该方法的核心,它结合基于传感器的障碍物观测来优化决策。这种设计在保持适应性和强大功能的同时,将计算需求降至最低。在NVIDIA Isaac Sim中使用ANYbotics四足机器人进行的模拟试验显示出很高的操纵精度,在高达10米的物体目标距离上平均定位误差为11厘米。此外,RL框架有效地将复杂环境中的路径规划集成在一起,实现了节能和稳定的操作。这些发现确立了该框架作为一种有前途的方法,适用于需要多功能性、效率和实际可部署性的先进机器人技术。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/ddcdf3527402/sensors-25-01565-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/d1d2ec83679b/sensors-25-01565-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/0b6136874701/sensors-25-01565-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/603048ee239e/sensors-25-01565-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/c1f63e2777ee/sensors-25-01565-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/ceb47e5d78ea/sensors-25-01565-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/ddcdf3527402/sensors-25-01565-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/d1d2ec83679b/sensors-25-01565-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/0b6136874701/sensors-25-01565-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/603048ee239e/sensors-25-01565-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/c1f63e2777ee/sensors-25-01565-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/ceb47e5d78ea/sensors-25-01565-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/54ce/11902496/ddcdf3527402/sensors-25-01565-g006.jpg

相似文献

1
Hierarchical Reinforcement Learning for Quadrupedal Robots: Efficient Object Manipulation in Constrained Environments.四足机器人的分层强化学习:受限环境中的高效物体操纵
Sensors (Basel). 2025 Mar 4;25(5):1565. doi: 10.3390/s25051565.
2
Review of Learning-Based Robotic Manipulation in Cluttered Environments.基于学习的杂乱环境机器人操作综述。
Sensors (Basel). 2022 Oct 18;22(20):7938. doi: 10.3390/s22207938.
3
Sample-efficient and occlusion-robust reinforcement learning for robotic manipulation via multimodal fusion dualization and representation normalization.
Neural Netw. 2025 May;185:107202. doi: 10.1016/j.neunet.2025.107202. Epub 2025 Jan 30.
4
RL-QPSO net: deep reinforcement learning-enhanced QPSO for efficient mobile robot path planning.RL-QPSO网络:用于高效移动机器人路径规划的深度强化学习增强型量子粒子群优化算法
Front Neurorobot. 2025 Jan 8;18:1464572. doi: 10.3389/fnbot.2024.1464572. eCollection 2024.
5
Learning robust autonomous navigation and locomotion for wheeled-legged robots.学习轮式腿式机器人的鲁棒自主导航与运动。
Sci Robot. 2024 Apr 24;9(89):eadi9641. doi: 10.1126/scirobotics.adi9641.
6
Learning-based control approaches for service robots on cloth manipulation and dressing assistance: a comprehensive review.基于学习的服务机器人布料操作和穿衣辅助控制方法:全面综述。
J Neuroeng Rehabil. 2022 Nov 3;19(1):117. doi: 10.1186/s12984-022-01078-4.
7
Diffusion Model-Based Path Follower for a Salamander-Like Robot.基于扩散模型的类蝾螈机器人路径跟踪器
IEEE Trans Neural Netw Learn Syst. 2025 Aug;36(8):14399-14413. doi: 10.1109/TNNLS.2025.3549307.
8
Novel deep reinforcement learning based collision avoidance approach for path planning of robots in unknown environment.基于新型深度强化学习的未知环境中机器人路径规划碰撞避免方法。
PLoS One. 2025 Jan 16;20(1):e0312559. doi: 10.1371/journal.pone.0312559. eCollection 2025.
9
Sim-to-real latent prediction: Transferring visual non-prehensile manipulation policies.从模拟到现实的潜在预测:转移视觉非抓取式操纵策略。
Front Robot AI. 2023 Jan 12;9:1067502. doi: 10.3389/frobt.2022.1067502. eCollection 2022.
10
Human-to-Robot Handover Based on Reinforcement Learning.基于强化学习的人机交互。
Sensors (Basel). 2024 Sep 27;24(19):6275. doi: 10.3390/s24196275.

本文引用的文献

1
ANYmal parkour: Learning agile navigation for quadrupedal robots.ANYmal 跑酷:学习四足机器人的敏捷导航。
Sci Robot. 2024 Mar 13;9(88):eadi7566. doi: 10.1126/scirobotics.adi7566.