• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

DPODv2:基于密集对应关系的 6 自由度位姿估计。

DPODv2: Dense Correspondence-Based 6 DoF Pose Estimation.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):7417-7435. doi: 10.1109/TPAMI.2021.3118833. Epub 2022 Oct 4.

DOI:10.1109/TPAMI.2021.3118833
PMID:34623263
Abstract

We propose a three-stage 6 DoF object detection method called DPODv2 (Dense Pose Object Detector) that relies on dense correspondences. We combine a 2D object detector with a dense correspondence estimation network and a multi-view pose refinement method to estimate a full 6 DoF pose. Unlike other deep learning methods that are typically restricted to monocular RGB images, we propose a unified deep learning network allowing different imaging modalities to be used (RGB or Depth). Moreover, we propose a novel pose refinement method, that is based on differentiable rendering. The main concept is to compare predicted and rendered correspondences in multiple views to obtain a pose which is consistent with predicted correspondences in all views. Our proposed method is evaluated rigorously on different data modalities and types of training data in a controlled setup. The main conclusions is that RGB excels in correspondence estimation, while depth contributes to the pose accuracy if good 3D-3D correspondences are available. Naturally, their combination achieves the overall best performance. We perform an extensive evaluation and an ablation study to analyze and validate the results on several challenging datasets. DPODv2 achieves excellent results on all of them while still remaining fast and scalable independent of the used data modality and the type of training data.

摘要

我们提出了一种称为 DPODv2(密集姿态对象检测)的三阶段 6DoF 对象检测方法,该方法依赖于密集对应关系。我们将 2D 对象检测器与密集对应估计网络和多视图姿态细化方法相结合,以估计完整的 6DoF 姿态。与其他通常仅限于单目 RGB 图像的深度学习方法不同,我们提出了一种统一的深度学习网络,允许使用不同的成像模式(RGB 或深度)。此外,我们提出了一种新颖的姿态细化方法,该方法基于可微分渲染。主要思想是在多个视图中比较预测的和渲染的对应关系,以获得与所有视图中预测的对应关系一致的姿态。我们在受控设置下,使用不同的数据模态和类型的训练数据对提出的方法进行了严格的评估。主要结论是 RGB 在对应估计方面表现出色,而深度如果有良好的 3D-3D 对应关系,则有助于提高姿态准确性。自然地,它们的组合实现了整体最佳性能。我们进行了广泛的评估和消融研究,以在几个具有挑战性的数据集上分析和验证结果。DPODv2 在所有这些数据集上都取得了优异的结果,同时仍然保持快速和可扩展,与所使用的数据模态和训练数据的类型无关。

相似文献

1
DPODv2: Dense Correspondence-Based 6 DoF Pose Estimation.DPODv2:基于密集对应关系的 6 自由度位姿估计。
IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):7417-7435. doi: 10.1109/TPAMI.2021.3118833. Epub 2022 Oct 4.
2
RNNPose: 6-DoF Object Pose Estimation via Recurrent Correspondence Field Estimation and Pose Optimization.RNNPose:通过循环对应场估计和姿态优化实现的6自由度物体姿态估计
IEEE Trans Pattern Anal Mach Intell. 2024 Jul;46(7):4669-4683. doi: 10.1109/TPAMI.2024.3360181. Epub 2024 Jun 5.
3
EPro-PnP: Generalized End-to-End Probabilistic Perspective-N-Points for Monocular Object Pose Estimation.EPro-PnP:用于单目物体位姿估计的广义端到端概率视角N点法
IEEE Trans Pattern Anal Mach Intell. 2024 Jan 16;PP. doi: 10.1109/TPAMI.2024.3354997.
4
DeepHMap++: Combined Projection Grouping and Correspondence Learning for Full DoF Pose Estimation.DeepHMap++:用于全自由度姿态估计的组合投影分组与对应学习
Sensors (Basel). 2019 Feb 28;19(5):1032. doi: 10.3390/s19051032.
5
Top-Down System for Multi-Person 3D Absolute Pose Estimation from Monocular Videos.基于单目视频的多人 3D 绝对姿态估计的自顶向下系统。
Sensors (Basel). 2022 May 28;22(11):4109. doi: 10.3390/s22114109.
6
Fast and Robust Multi-Person 3D Pose Estimation and Tracking From Multiple Views.快速稳健的多人三维姿态估计与多视角跟踪。
IEEE Trans Pattern Anal Mach Intell. 2022 Oct;44(10):6981-6992. doi: 10.1109/TPAMI.2021.3098052. Epub 2022 Sep 14.
7
Coarse-to-Fine Hand-Object Pose Estimation with Interaction-Aware Graph Convolutional Network.基于交互感知图卷积网络的粗到细手物姿态估计。
Sensors (Basel). 2021 Dec 3;21(23):8092. doi: 10.3390/s21238092.
8
Learning 3D Human Shape and Pose From Dense Body Parts.从密集的身体部位学习 3D 人体形状和姿势。
IEEE Trans Pattern Anal Mach Intell. 2022 May;44(5):2610-2627. doi: 10.1109/TPAMI.2020.3042341. Epub 2022 Apr 1.
9
Towards markerless surgical tool and hand pose estimation.面向无标记手术工具和手部姿势估计。
Int J Comput Assist Radiol Surg. 2021 May;16(5):799-808. doi: 10.1007/s11548-021-02369-2. Epub 2021 Apr 21.
10
Visual Camera Re-Localization From RGB and RGB-D Images Using DSAC.基于 DSAC 的 RGB 和 RGB-D 图像的视觉相机重定位
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):5847-5865. doi: 10.1109/TPAMI.2021.3070754. Epub 2022 Aug 4.

引用本文的文献

1
A Survey of 6DoF Object Pose Estimation Methods for Different Application Scenarios.针对不同应用场景的六自由度物体位姿估计方法综述
Sensors (Basel). 2024 Feb 7;24(4):1076. doi: 10.3390/s24041076.
2
A Survey of 6D Object Detection Based on 3D Models for Industrial Applications.基于3D模型的工业应用6D目标检测研究
J Imaging. 2022 Feb 24;8(3):53. doi: 10.3390/jimaging8030053.