• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于多层次局部特征和全局特征的回归相机位姿估计。

Regression-Based Camera Pose Estimation through Multi-Level Local Features and Global Features.

机构信息

School of Electronic Engineering and Computer Science, Queen Mary University of London, London E1 4NS, UK.

School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China.

出版信息

Sensors (Basel). 2023 Apr 18;23(8):4063. doi: 10.3390/s23084063.

DOI:10.3390/s23084063
PMID:37112404
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10146233/
Abstract

Accurate and robust camera pose estimation is essential for high-level applications such as augmented reality and autonomous driving. Despite the development of global feature-based camera pose regression methods and local feature-based matching guided pose estimation methods, challenging conditions, such as illumination changes and viewpoint changes, as well as inaccurate keypoint localization, continue to affect the performance of camera pose estimation. In this paper, we propose a novel relative camera pose regression framework that uses global features with rotation consistency and local features with rotation invariance. First, we apply a multi-level deformable network to detect and describe local features, which can learn appearances and gradient information sensitive to rotation variants. Second, we process the detection and description processes using the results from pixel correspondences of the input image pairs. Finally, we propose a novel loss that combines relative regression loss and absolute regression loss, incorporating global features with geometric constraints to optimize the pose estimation model. Our extensive experiments report satisfactory accuracy on the 7Scenes dataset with an average mean translation error of 0.18 m and a rotation error of 7.44° using image pairs as input. Ablation studies were also conducted to verify the effectiveness of the proposed method in the tasks of pose estimation and image matching using the 7Scenes and HPatches datasets.

摘要

准确而稳健的相机位姿估计对于高级应用(如增强现实和自动驾驶)至关重要。尽管已经开发了基于全局特征的相机位姿回归方法和基于局部特征的匹配引导位姿估计方法,但挑战性的条件,如光照变化和视角变化,以及不准确的关键点定位,仍然会影响相机位姿估计的性能。在本文中,我们提出了一种新的相对相机位姿回归框架,该框架使用具有旋转一致性的全局特征和具有旋转不变性的局部特征。首先,我们应用多层可变形网络来检测和描述局部特征,该特征可以学习对旋转变体敏感的外观和梯度信息。其次,我们使用输入图像对的像素对应关系的结果来处理检测和描述过程。最后,我们提出了一种新的损失函数,它结合了相对回归损失和绝对回归损失,将具有几何约束的全局特征纳入其中,以优化位姿估计模型。我们在 7Scenes 数据集上进行了广泛的实验,报告了令人满意的精度,平均平移误差为 0.18 米,旋转误差为 7.44°,输入的是图像对。我们还进行了消融研究,以验证该方法在使用 7Scenes 和 HPatches 数据集进行位姿估计和图像匹配任务中的有效性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/b0bd95c665f1/sensors-23-04063-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/d2531768a5af/sensors-23-04063-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/67c8207d4cb0/sensors-23-04063-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/0beea28fa345/sensors-23-04063-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/a46adaf6398e/sensors-23-04063-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/4ccdb1c9b631/sensors-23-04063-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/3aaca5caf571/sensors-23-04063-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/1292500c2969/sensors-23-04063-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/9d6ad889667a/sensors-23-04063-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/f667b3d47775/sensors-23-04063-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/d24e901f0a58/sensors-23-04063-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/ce782b5d10f9/sensors-23-04063-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/b0bd95c665f1/sensors-23-04063-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/d2531768a5af/sensors-23-04063-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/67c8207d4cb0/sensors-23-04063-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/0beea28fa345/sensors-23-04063-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/a46adaf6398e/sensors-23-04063-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/4ccdb1c9b631/sensors-23-04063-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/3aaca5caf571/sensors-23-04063-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/1292500c2969/sensors-23-04063-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/9d6ad889667a/sensors-23-04063-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/f667b3d47775/sensors-23-04063-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/d24e901f0a58/sensors-23-04063-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/ce782b5d10f9/sensors-23-04063-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/14c5/10146233/b0bd95c665f1/sensors-23-04063-g012.jpg

相似文献

1
Regression-Based Camera Pose Estimation through Multi-Level Local Features and Global Features.基于多层次局部特征和全局特征的回归相机位姿估计。
Sensors (Basel). 2023 Apr 18;23(8):4063. doi: 10.3390/s23084063.
2
Learning Geometric Feature Embedding with Transformers for Image Matching.基于 Transformer 的图像匹配的几何特征嵌入学习。
Sensors (Basel). 2022 Dec 15;22(24):9882. doi: 10.3390/s22249882.
3
Image partitioning and illumination in image-based pose detection for teleoperated flexible endoscopes.基于图像的远程操作柔性内窥镜位姿检测中的图像分区和光照。
Artif Intell Med. 2013 Nov;59(3):185-96. doi: 10.1016/j.artmed.2013.09.002. Epub 2013 Oct 10.
4
Multi-level feature fusion and joint refinement for simultaneous object pose estimation and camera localization.用于同时进行目标位姿估计和相机定位的多层次特征融合和联合细化。
Neural Netw. 2024 Jun;174:106238. doi: 10.1016/j.neunet.2024.106238. Epub 2024 Mar 16.
5
Detection, segmentation, and 3D pose estimation of surgical tools using convolutional neural networks and algebraic geometry.使用卷积神经网络和代数几何进行手术工具的检测、分割和三维姿态估计。
Med Image Anal. 2021 May;70:101994. doi: 10.1016/j.media.2021.101994. Epub 2021 Feb 7.
6
Unsupervised Monocular Depth and Camera Pose Estimation with Multiple Masks and Geometric Consistency Constraints.基于多掩码和几何一致性约束的无监督单目深度与相机位姿估计
Sensors (Basel). 2023 Jun 4;23(11):5329. doi: 10.3390/s23115329.
7
Dynamic Keypoint Detection Network for Image Matching.用于图像匹配的动态关键点检测网络
IEEE Trans Pattern Anal Mach Intell. 2023 Dec;45(12):14404-14419. doi: 10.1109/TPAMI.2023.3307889. Epub 2023 Nov 3.
8
6-DOF Pose Estimation of a Robotic Navigation Aid by Tracking Visual and Geometric Features.通过跟踪视觉和几何特征实现机器人导航辅助设备的六自由度姿态估计
IEEE Trans Autom Sci Eng. 2015 Oct;12(4):1169-1180. doi: 10.1109/TASE.2015.2469726. Epub 2015 Oct 5.
9
A lightweight Yunnan Xiaomila detection and pose estimation based on improved YOLOv8.一种基于改进YOLOv8的轻量化云南小米辣检测与姿态估计
Front Plant Sci. 2024 Jun 5;15:1421381. doi: 10.3389/fpls.2024.1421381. eCollection 2024.
10
3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data.3D 仿射:一种使用 RGB-D 传感器数据进行视点不变性的局部图像特征嵌入。
Sensors (Basel). 2019 Jan 12;19(2):291. doi: 10.3390/s19020291.

本文引用的文献

1
Deep Learning Framework for Controlling Work Sequence in Collaborative Human-Robot Assembly Processes.深度学习框架在协作式人机装配过程中控制作业序列。
Sensors (Basel). 2023 Jan 3;23(1):553. doi: 10.3390/s23010553.
2
Estimating Ground Reaction Forces from Two-Dimensional Pose Data: A Biomechanics-Based Comparison of AlphaPose, BlazePose, and OpenPose.基于二维姿态数据估计地面反作用力:AlphaPose、BlazePose 和 OpenPose 的生物力学比较。
Sensors (Basel). 2022 Dec 21;23(1):78. doi: 10.3390/s23010078.