• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于条件变分自编码器的动态运动用于多任务模仿学习。

Conditional variational auto encoder based dynamic motion for multitask imitation learning.

作者信息

Xu Binzhao, Ud Din Muhayy, Hussain Irfan

机构信息

Khalifa University Center for Autonomous Robotic Systems (KUCARS), Khalifa University of Science and Technology, Abu Dhabi, UAE.

出版信息

Sci Rep. 2025 Mar 17;15(1):9196. doi: 10.1038/s41598-025-93888-4.

DOI:10.1038/s41598-025-93888-4
PMID:40097597
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11914600/
Abstract

The dynamic motion primitive-based (DMP) method is effective for learning from demonstrations. However, most current DMP-based methods focus on learning one task with one module. Although, some deep learning based frameworks can learn multi-task simultaneously. However, these methods require a large amount of training data and have limited generalization of the learned behavior to untrained states. In this paper, we propose a framework that combines the advantages of the traditional DMP-based method and conditional variational auto-encoder (cVAE). The encoder and decoder comprise a dynamic system and a deep neural network. Instead of generating a trajectory directly, deep neural networks are used to generate torque conditioned on the task parameters. This torque is then used to produce the desired trajectory in the dynamic system, based on the final state. In this way, the generated trajectory can adapt to the new goal position, similar to DMP. We also propose a fine-tuning method to guarantee the via-point constraint. Our model is trained and tested on the handwritten digit number dataset and robotic manipulation tasks, such as pushing, reaching, and grasping. Finally, the proposed model is also validated in a real robotic environment with a UR10 manipulator. Compared to traditional data-demanding deep learning-based methods, it is remarkable that our proposed method can achieve a 100% success rate in the reaching task and a 93.33% success rate in pushing and grasping tasks, with only one demonstration provided for each task.

摘要

基于动态运动基元(DMP)的方法在从示范中学习方面很有效。然而,当前大多数基于DMP的方法都专注于用一个模块学习一项任务。虽然,一些基于深度学习的框架可以同时学习多任务。然而,这些方法需要大量的训练数据,并且对未训练状态下学习到的行为的泛化能力有限。在本文中,我们提出了一个结合传统基于DMP的方法和条件变分自编码器(cVAE)优点的框架。编码器和解码器由一个动态系统和一个深度神经网络组成。深度神经网络不是直接生成轨迹,而是用于根据任务参数生成扭矩。然后,基于最终状态,该扭矩用于在动态系统中产生所需的轨迹。通过这种方式,生成的轨迹可以像DMP一样适应新的目标位置。我们还提出了一种微调方法来保证过点约束。我们的模型在手写数字数据集和机器人操作任务(如推、伸手和抓握)上进行了训练和测试。最后,所提出的模型也在配备UR10机械手的真实机器人环境中得到了验证。与传统的基于深度学习且需要大量数据的方法相比,值得注意的是,我们提出的方法在伸手任务中可以达到100%的成功率,在推和抓握任务中可以达到93.33%的成功率,每个任务仅提供一次示范。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/1161f710189a/41598_2025_93888_Fig19_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/61ba3e445377/41598_2025_93888_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/bf0fee730b51/41598_2025_93888_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/78801343da5c/41598_2025_93888_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/2bc5ec0e31c8/41598_2025_93888_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/d7bb5ef1dd41/41598_2025_93888_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/ca09401c1ed0/41598_2025_93888_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/d25d8aa11efe/41598_2025_93888_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/656d9607e9bd/41598_2025_93888_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/f495e3253f0a/41598_2025_93888_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/ed061d0876c6/41598_2025_93888_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/21e905b3764a/41598_2025_93888_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/c807d61e3edf/41598_2025_93888_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/b859a01bf684/41598_2025_93888_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/11061a97e1fa/41598_2025_93888_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/7e4ee39b2c59/41598_2025_93888_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/e23da29a3481/41598_2025_93888_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/4baceeeb5d1e/41598_2025_93888_Fig17_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/a304658641bc/41598_2025_93888_Fig18_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/1161f710189a/41598_2025_93888_Fig19_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/61ba3e445377/41598_2025_93888_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/bf0fee730b51/41598_2025_93888_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/78801343da5c/41598_2025_93888_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/2bc5ec0e31c8/41598_2025_93888_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/d7bb5ef1dd41/41598_2025_93888_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/ca09401c1ed0/41598_2025_93888_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/d25d8aa11efe/41598_2025_93888_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/656d9607e9bd/41598_2025_93888_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/f495e3253f0a/41598_2025_93888_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/ed061d0876c6/41598_2025_93888_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/21e905b3764a/41598_2025_93888_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/c807d61e3edf/41598_2025_93888_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/b859a01bf684/41598_2025_93888_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/11061a97e1fa/41598_2025_93888_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/7e4ee39b2c59/41598_2025_93888_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/e23da29a3481/41598_2025_93888_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/4baceeeb5d1e/41598_2025_93888_Fig17_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/a304658641bc/41598_2025_93888_Fig18_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ea54/11914600/1161f710189a/41598_2025_93888_Fig19_HTML.jpg

相似文献

1
Conditional variational auto encoder based dynamic motion for multitask imitation learning.基于条件变分自编码器的动态运动用于多任务模仿学习。
Sci Rep. 2025 Mar 17;15(1):9196. doi: 10.1038/s41598-025-93888-4.
2
Posterior estimation using deep learning: a simulation study of compartmental modeling in dynamic positron emission tomography.深度学习的后验估计:动态正电子发射断层扫描中隔室建模的模拟研究。
Med Phys. 2023 Mar;50(3):1539-1548. doi: 10.1002/mp.16078. Epub 2022 Nov 18.
3
Posterior Estimation Using Deep Learning: A Simulation Study of Compartmental Modeling in Dynamic PET.使用深度学习的后验估计:动态PET中房室模型的模拟研究
ArXiv. 2023 Mar 17:arXiv:2303.10057v1.
4
Peg-in-hole assembly skill imitation learning method based on ProMPs under task geometric representation.基于任务几何表示的ProMPs的插销入孔装配技能模仿学习方法
Front Neurorobot. 2023 Nov 9;17:1320251. doi: 10.3389/fnbot.2023.1320251. eCollection 2023.
5
Supervised Multi-Layer Conditional Variational Auto-Encoder for Process Modeling and Soft Sensor.用于过程建模和软传感器的监督式多层条件变分自编码器
Sensors (Basel). 2023 Nov 14;23(22):9175. doi: 10.3390/s23229175.
6
Modeling and augmenting of fMRI data using deep recurrent variational auto-encoder.使用深度递归变分自动编码器对 fMRI 数据进行建模和增强。
J Neural Eng. 2021 Jul 23;18(4). doi: 10.1088/1741-2552/ac1179.
7
A trajectory outlier detection method based on variational auto-encoder.一种基于变分自编码器的轨迹异常检测方法。
Math Biosci Eng. 2023 Jul 14;20(8):15075-15093. doi: 10.3934/mbe.2023675.
8
Imitation and mirror systems in robots through Deep Modality Blending Networks.机器人中的模仿和镜像系统通过深度模态混合网络。
Neural Netw. 2022 Feb;146:22-35. doi: 10.1016/j.neunet.2021.11.004. Epub 2021 Nov 16.
9
Deep imitation learning for 3D navigation tasks.用于3D导航任务的深度模仿学习
Neural Comput Appl. 2018;29(7):389-404. doi: 10.1007/s00521-017-3241-z. Epub 2017 Dec 4.
10
Behavior policy learning: Learning multi-stage tasks solution sketches and model-based controllers.行为策略学习:学习多阶段任务解决方案草图和基于模型的控制器。
Front Robot AI. 2022 Oct 12;9:974537. doi: 10.3389/frobt.2022.974537. eCollection 2022.

本文引用的文献

1
A Survey of Imitation Learning: Algorithms, Recent Developments, and Challenges.模仿学习综述:算法、最新进展与挑战
IEEE Trans Cybern. 2024 Dec;54(12):7173-7186. doi: 10.1109/TCYB.2024.3395626. Epub 2024 Nov 27.
2
Soft robotic manipulator for intraoperative MRI-guided transoral laser microsurgery.用于术中 MRI 引导经口激光微创手术的软体机器人操作臂。
Sci Robot. 2021 Aug 18;6(57). doi: 10.1126/scirobotics.abg5575.
3
Dynamical movement primitives: learning attractor models for motor behaviors.动力运动基元:学习运动行为的吸引子模型。
Neural Comput. 2013 Feb;25(2):328-73. doi: 10.1162/NECO_a_00393. Epub 2012 Nov 13.