• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

DeepHMap++:用于全自由度姿态估计的组合投影分组与对应学习

DeepHMap++: Combined Projection Grouping and Correspondence Learning for Full DoF Pose Estimation.

机构信息

State Key Laboratory of Robotics, Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang 110016, China.

Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110016, China.

出版信息

Sensors (Basel). 2019 Feb 28;19(5):1032. doi: 10.3390/s19051032.

DOI:10.3390/s19051032
PMID:30823451
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC6427142/
Abstract

In recent years, estimating the 6D pose of object instances with convolutional neural network (CNN) has received considerable attention. Depending on whether intermediate cues are used, the relevant literature can be roughly divided into two broad categories: direct methods and two-stage pipelines. For the latter, intermediate cues, such as 3D object coordinates, semantic keypoints, or virtual control points instead of pose parameters are regressed by CNN in the first stage. Object pose can then be solved by correspondence constraints constructed with these intermediate cues. In this paper, we focus on the postprocessing of a two-stage pipeline and propose to combine two learning concepts for estimating object pose under challenging scenes: projection grouping on one side, and correspondence learning on the other. We firstly employ a local-patch based method to predict projection heatmaps which denote the confidence distribution of projection of 3D bounding box's corners. A projection grouping module is then proposed to remove redundant local maxima from each layer of heatmaps. Instead of directly feeding 2D⁻3D correspondences to the perspective-n-point (PnP) algorithm, multiple correspondence hypotheses are sampled from local maxima and its corresponding neighborhood and ranked by a correspondence⁻evaluation network. Finally, correspondences with higher confidence are selected to determine object pose. Extensive experiments on three public datasets demonstrate that the proposed framework outperforms several state of the art methods.

摘要

近年来,使用卷积神经网络 (CNN) 估计物体实例的 6D 姿态受到了广泛关注。根据是否使用中间线索,可以将相关文献大致分为两类:直接方法和两阶段管道。对于后者,中间线索(如 3D 物体坐标、语义关键点或虚拟控制点)而不是姿态参数,由 CNN 在第一阶段回归。然后可以通过使用这些中间线索构建的对应关系约束来解决物体姿态。在本文中,我们专注于两阶段管道的后处理,并提出了两种用于在具有挑战性的场景下估计物体姿态的学习概念:一方面是投影分组,另一方面是对应学习。我们首先采用基于局部补丁的方法来预测投影热力图,该热力图表示 3D 边界框角的投影置信度分布。然后提出了一个投影分组模块,从每个热力图层中去除冗余的局部极大值。我们不是直接将 2D-3D 对应关系馈送到透视 N 点 (PnP) 算法中,而是从局部极大值及其对应邻域中采样多个对应假设,并由对应评估网络对其进行排序。最后,选择置信度更高的对应关系来确定物体姿态。在三个公共数据集上的广泛实验表明,所提出的框架优于几种最先进的方法。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/77736f89ada3/sensors-19-01032-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/8f5f731bc195/sensors-19-01032-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/f6b24632407f/sensors-19-01032-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/e4610590f362/sensors-19-01032-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/45992f8abb83/sensors-19-01032-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/11d0bc25d9b2/sensors-19-01032-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/8888354070c6/sensors-19-01032-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/89b4aad854c0/sensors-19-01032-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/b6f8e036dbf4/sensors-19-01032-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/0584e13fccf0/sensors-19-01032-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/77736f89ada3/sensors-19-01032-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/8f5f731bc195/sensors-19-01032-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/f6b24632407f/sensors-19-01032-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/e4610590f362/sensors-19-01032-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/45992f8abb83/sensors-19-01032-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/11d0bc25d9b2/sensors-19-01032-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/8888354070c6/sensors-19-01032-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/89b4aad854c0/sensors-19-01032-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/b6f8e036dbf4/sensors-19-01032-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/0584e13fccf0/sensors-19-01032-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3970/6427142/77736f89ada3/sensors-19-01032-g010.jpg

相似文献

1
DeepHMap++: Combined Projection Grouping and Correspondence Learning for Full DoF Pose Estimation.DeepHMap++:用于全自由度姿态估计的组合投影分组与对应学习
Sensors (Basel). 2019 Feb 28;19(5):1032. doi: 10.3390/s19051032.
2
RNNPose: 6-DoF Object Pose Estimation via Recurrent Correspondence Field Estimation and Pose Optimization.RNNPose:通过循环对应场估计和姿态优化实现的6自由度物体姿态估计
IEEE Trans Pattern Anal Mach Intell. 2024 Jul;46(7):4669-4683. doi: 10.1109/TPAMI.2024.3360181. Epub 2024 Jun 5.
3
EPro-PnP: Generalized End-to-End Probabilistic Perspective-N-Points for Monocular Object Pose Estimation.EPro-PnP:用于单目物体位姿估计的广义端到端概率视角N点法
IEEE Trans Pattern Anal Mach Intell. 2024 Jan 16;PP. doi: 10.1109/TPAMI.2024.3354997.
4
DPODv2: Dense Correspondence-Based 6 DoF Pose Estimation.DPODv2:基于密集对应关系的 6 自由度位姿估计。
IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):7417-7435. doi: 10.1109/TPAMI.2021.3118833. Epub 2022 Oct 4.
5
Globally-Optimal Inlier Set Maximisation for Camera Pose and Correspondence Estimation.用于相机姿态和对应估计的全局最优内点集最大化
IEEE Trans Pattern Anal Mach Intell. 2020 Feb;42(2):328-342. doi: 10.1109/TPAMI.2018.2848650. Epub 2018 Jun 19.
6
SynPo-Net-Accurate and Fast CNN-Based 6DoF Object Pose Estimation Using Synthetic Training.基于 SynPo-Net 的准确、快速的 CNN 六自由度物体位姿估计方法,使用合成训练。
Sensors (Basel). 2021 Jan 5;21(1):300. doi: 10.3390/s21010300.
7
Coarse-to-Fine Hand-Object Pose Estimation with Interaction-Aware Graph Convolutional Network.基于交互感知图卷积网络的粗到细手物姿态估计。
Sensors (Basel). 2021 Dec 3;21(23):8092. doi: 10.3390/s21238092.
8
Learning Two-View Correspondences and Geometry via Local Neighborhood Correlation.通过局部邻域相关性学习双视图对应关系和几何结构。
Entropy (Basel). 2021 Aug 9;23(8):1024. doi: 10.3390/e23081024.
9
Real-Time 3D Hand Pose Estimation with 3D Convolutional Neural Networks.基于3D卷积神经网络的实时3D手部姿态估计
IEEE Trans Pattern Anal Mach Intell. 2019 Apr;41(4):956-970. doi: 10.1109/TPAMI.2018.2827052. Epub 2018 Apr 16.
10
PVNet: Pixel-Wise Voting Network for 6DoF Object Pose Estimation.PVNet:用于 6DoF 对象位姿估计的像素级投票网络。
IEEE Trans Pattern Anal Mach Intell. 2022 Jun;44(6):3212-3223. doi: 10.1109/TPAMI.2020.3047388. Epub 2022 May 5.

引用本文的文献

1
A Survey of 6DoF Object Pose Estimation Methods for Different Application Scenarios.针对不同应用场景的六自由度物体位姿估计方法综述
Sensors (Basel). 2024 Feb 7;24(4):1076. doi: 10.3390/s24041076.
2
RGB-D-Based Pose Estimation of Workpieces with Semantic Segmentation and Point Cloud Registration.基于RGB-D的工件姿态估计:语义分割与点云配准
Sensors (Basel). 2019 Apr 19;19(8):1873. doi: 10.3390/s19081873.

本文引用的文献

1
Faceness-Net: Face Detection through Deep Facial Part Responses.Faceness-Net:基于深度人脸部件响应的人脸检测。
IEEE Trans Pattern Anal Mach Intell. 2018 Aug;40(8):1845-1859. doi: 10.1109/TPAMI.2017.2738644. Epub 2017 Aug 11.
2
Robust 3D Object Tracking from Monocular Images Using Stable Parts.基于稳定部件的单目图像鲁棒 3D 目标跟踪
IEEE Trans Pattern Anal Mach Intell. 2018 Jun;40(6):1465-1479. doi: 10.1109/TPAMI.2017.2708711. Epub 2017 May 26.
3
Latent-Class Hough Forests for 6 DoF Object Pose Estimation.潜在类别霍夫森林用于 6 自由度物体位姿估计。
IEEE Trans Pattern Anal Mach Intell. 2018 Jan;40(1):119-132. doi: 10.1109/TPAMI.2017.2665623. Epub 2017 Feb 7.