• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于机器人抓取检测的基于形状变形的6-D类别级物体位姿估计

Category-Level 6-D Object Pose Estimation With Shape Deformation for Robotic Grasp Detection.

作者信息

Yu Sheng, Zhai Di-Hua, Guan Yuyin, Xia Yuanqing

出版信息

IEEE Trans Neural Netw Learn Syst. 2025 Jan;36(1):1857-1871. doi: 10.1109/TNNLS.2023.3330011. Epub 2025 Jan 7.

DOI:10.1109/TNNLS.2023.3330011
PMID:37962999
Abstract

Category-level 6-D object pose estimation plays a crucial role in achieving reliable robotic grasp detection. However, the disparity between synthetic and real datasets hinders the direct transfer of models trained on synthetic data to real-world scenarios, leading to ineffective results. Additionally, creating large-scale real datasets is a time-consuming and labor-intensive task. To overcome these challenges, we propose CatDeform, a novel category-level object pose estimation network trained on synthetic data but capable of delivering good performance on real datasets. In our approach, we introduce a transformer-based fusion module that enables the network to leverage multiple sources of information and enhance prediction accuracy through feature fusion. To ensure proper deformation of the prior point cloud to align with scene objects, we propose a transformer-based attention module that deforms the prior point cloud from both geometric and feature perspectives. Building upon CatDeform, we design a two-branch network for supervised learning, bridging the gap between synthetic and real datasets and achieving high-precision pose estimation in real-world scenes using predominantly synthetic data supplemented with a small amount of real data. To minimize reliance on large-scale real datasets, we train the network in a self-supervised manner by estimating object poses in real scenes based on the synthetic dataset without manual annotation. We conduct training and testing on CAMERA25 and REAL275 datasets, and our experimental results demonstrate that the proposed method outperforms state-of-the-art (SOTA) techniques in both self-supervised and supervised training paradigms. Finally, we apply CatDeform to object pose estimation and robotic grasp experiments in real-world scenarios, showcasing a higher grasp success rate.

摘要

类别级6D物体位姿估计在实现可靠的机器人抓取检测中起着至关重要的作用。然而,合成数据集和真实数据集之间的差异阻碍了在合成数据上训练的模型直接应用于现实世界场景,导致效果不佳。此外,创建大规模真实数据集是一项耗时且费力的任务。为了克服这些挑战,我们提出了CatDeform,这是一种新颖的类别级物体位姿估计网络,它在合成数据上进行训练,但能够在真实数据集上表现出良好的性能。在我们的方法中,我们引入了一个基于Transformer的融合模块,使网络能够利用多种信息源,并通过特征融合提高预测精度。为了确保先验点云的适当变形以与场景物体对齐,我们提出了一个基于Transformer的注意力模块,从几何和特征两个角度对先验点云进行变形。基于CatDeform,我们设计了一个用于监督学习的双分支网络,弥合了合成数据集和真实数据集之间的差距,并使用主要是合成数据并辅以少量真实数据在现实世界场景中实现高精度位姿估计。为了尽量减少对大规模真实数据集的依赖,我们通过基于合成数据集在无人工标注的情况下估计真实场景中的物体位姿,以自监督的方式训练网络。我们在CAMERA25和REAL275数据集上进行训练和测试,实验结果表明,所提出的方法在自监督和监督训练范式中均优于现有技术(SOTA)。最后,我们将CatDeform应用于现实世界场景中的物体位姿估计和机器人抓取实验,展示了更高的抓取成功率。

相似文献

1
Category-Level 6-D Object Pose Estimation With Shape Deformation for Robotic Grasp Detection.用于机器人抓取检测的基于形状变形的6-D类别级物体位姿估计
IEEE Trans Neural Netw Learn Syst. 2025 Jan;36(1):1857-1871. doi: 10.1109/TNNLS.2023.3330011. Epub 2025 Jan 7.
2
Multi-level feature fusion and joint refinement for simultaneous object pose estimation and camera localization.用于同时进行目标位姿估计和相机定位的多层次特征融合和联合细化。
Neural Netw. 2024 Jun;174:106238. doi: 10.1016/j.neunet.2024.106238. Epub 2024 Mar 16.
3
MH6D: Multi-Hypothesis Consistency Learning for Category-Level 6-D Object Pose Estimation.MH6D:用于类别级6D物体姿态估计的多假设一致性学习
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):4820-4833. doi: 10.1109/TNNLS.2024.3360712. Epub 2025 Feb 28.
4
MSSPA-GC: Multi-Scale Shape Prior Adaptation with 3D Graph Convolutions for Category-Level Object Pose Estimation.MSSPA-GC:基于 3D 图卷积的多尺度形状先验自适应的类别级物体位姿估计。
Neural Netw. 2023 Sep;166:609-621. doi: 10.1016/j.neunet.2023.07.037. Epub 2023 Jul 31.
5
6-D Object Pose Estimation Based on Point Pair Matching for Robotic Grasp Detection.基于点对匹配的6D物体姿态估计用于机器人抓取检测
IEEE Trans Neural Netw Learn Syst. 2025 Jul;36(7):11902-11916. doi: 10.1109/TNNLS.2024.3442433.
6
Diff9D: Diffusion-Based Domain-Generalized Category-Level 9-DoF Object Pose Estimation.Diff9D:基于扩散的领域通用类别级9自由度物体姿态估计
IEEE Trans Pattern Anal Mach Intell. 2025 Jul;47(7):5520-5537. doi: 10.1109/TPAMI.2025.3552132.
7
Instance-level 6D pose estimation based on multi-task parameter sharing for robotic grasping.基于多任务参数共享的实例级6D姿态估计用于机器人抓取。
Sci Rep. 2024 Apr 2;14(1):7801. doi: 10.1038/s41598-024-58590-x.
8
A neural learning approach for simultaneous object detection and grasp detection in cluttered scenes.一种用于在杂乱场景中同时进行目标检测和抓取检测的神经学习方法。
Front Comput Neurosci. 2023 Feb 20;17:1110889. doi: 10.3389/fncom.2023.1110889. eCollection 2023.
9
Category-Level Object Pose Estimation with Statistic Attention.基于统计注意力的类别级目标姿态估计
Sensors (Basel). 2024 Aug 19;24(16):5347. doi: 10.3390/s24165347.
10
6DoF assembly pose estimation dataset for robotic manipulation.用于机器人操作的6自由度装配姿态估计数据集。
Data Brief. 2024 Aug 14;56:110834. doi: 10.1016/j.dib.2024.110834. eCollection 2024 Oct.

引用本文的文献

1
Robot multi-target high performance grasping detection based on random sub-path fusion.基于随机子路径融合的机器人多目标高性能抓取检测
Sci Rep. 2025 Mar 13;15(1):8709. doi: 10.1038/s41598-025-93490-8.