• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

结合开放街道地图与卫星图像以增强跨视图地理定位

Combining OpenStreetMap with Satellite Imagery to Enhance Cross-View Geo-Localization.

作者信息

Hu Yuekun, Liu Yingfan, Hui Bin

机构信息

Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang 110016, China.

Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang 110016, China.

出版信息

Sensors (Basel). 2024 Dec 25;25(1):44. doi: 10.3390/s25010044.

DOI:10.3390/s25010044
PMID:39796834
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11723410/
Abstract

Cross-view geo-localization (CVGL) aims to determine the capture location of street-view images by matching them with corresponding 2D maps, such as satellite imagery. While recent bird's eye view (BEV)-based methods have advanced this task by addressing viewpoint and appearance differences, the existing approaches typically rely solely on either OpenStreetMap (OSM) data or satellite imagery, limiting localization robustness due to single-modality constraints. This paper presents a novel CVGL method that fuses OSM data with satellite imagery, leveraging their complementary strengths to enhance localization robustness. We integrate the semantic richness and structural information from OSM with the high-resolution visual details of satellite imagery, creating a unified 2D geospatial representation. Additionally, we employ a transformer-based BEV perception module that utilizes attention mechanisms to construct fine-grained BEV features from street-view images for matching with fused map features. Compared to state-of-the-art methods that utilize only OSM data, our approach achieves substantial improvements, with 12.05% and 12.06% recall enhancements on the KITTI benchmark for lateral and longitudinal localization within a 1-m error, respectively.

摘要

跨视角地理定位(CVGL)旨在通过将街景图像与相应的二维地图(如卫星图像)进行匹配来确定其拍摄位置。虽然最近基于鸟瞰图(BEV)的方法通过解决视角和外观差异推进了这项任务,但现有方法通常仅依赖于开放街道地图(OSM)数据或卫星图像,由于单模态约束而限制了定位的鲁棒性。本文提出了一种新颖的CVGL方法,该方法将OSM数据与卫星图像融合,利用它们的互补优势来增强定位鲁棒性。我们将来自OSM的语义丰富性和结构信息与卫星图像的高分辨率视觉细节相结合,创建一个统一的二维地理空间表示。此外,我们采用基于Transformer的BEV感知模块,该模块利用注意力机制从街景图像构建细粒度的BEV特征,以便与融合后的地图特征进行匹配。与仅使用OSM数据的现有方法相比,我们的方法取得了显著改进,在KITTI基准测试中,横向和纵向定位在1米误差范围内的召回率分别提高了12.05%和12.06%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/f127d8b6cd44/sensors-25-00044-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/065e88272874/sensors-25-00044-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/e5b2f9649076/sensors-25-00044-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/3a730a59305e/sensors-25-00044-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/6b4e1af7376a/sensors-25-00044-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/75715392b443/sensors-25-00044-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/e5d6479f0192/sensors-25-00044-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/76eeafc5b77b/sensors-25-00044-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/f127d8b6cd44/sensors-25-00044-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/065e88272874/sensors-25-00044-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/e5b2f9649076/sensors-25-00044-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/3a730a59305e/sensors-25-00044-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/6b4e1af7376a/sensors-25-00044-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/75715392b443/sensors-25-00044-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/e5d6479f0192/sensors-25-00044-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/76eeafc5b77b/sensors-25-00044-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d89f/11723410/f127d8b6cd44/sensors-25-00044-g008.jpg

相似文献

1
Combining OpenStreetMap with Satellite Imagery to Enhance Cross-View Geo-Localization.结合开放街道地图与卫星图像以增强跨视图地理定位
Sensors (Basel). 2024 Dec 25;25(1):44. doi: 10.3390/s25010044.
2
Geometry-Guided Street-View Panorama Synthesis From Satellite Imagery.基于卫星图像的几何引导街景全景合成
IEEE Trans Pattern Anal Mach Intell. 2022 Dec;44(12):10009-10022. doi: 10.1109/TPAMI.2022.3140750. Epub 2022 Nov 7.
3
UAV's Status Is Worth Considering: A Fusion Representations Matching Method for Geo-Localization.无人机的状态值得考虑:一种用于地理定位的融合表示匹配方法。
Sensors (Basel). 2023 Jan 8;23(2):720. doi: 10.3390/s23020720.
4
Unified and Real-Time Image Geo-Localization via Fine-Grained Overlap Estimation.通过细粒度重叠估计实现统一实时图像地理定位
IEEE Trans Image Process. 2024;33:5060-5072. doi: 10.1109/TIP.2024.3453008. Epub 2024 Sep 17.
5
Fast vehicle detection based on colored point cloud with bird's eye view representation.基于鸟瞰彩色点云的快速车辆检测。
Sci Rep. 2023 May 8;13(1):7447. doi: 10.1038/s41598-023-34479-z.
6
LiDAR-OSM-Based Vehicle Localization in GPS-Denied Environments by Using Constrained Particle Filter.基于激光雷达-开放街道地图的车辆在全球定位系统受限环境中的定位:使用约束粒子滤波器
Sensors (Basel). 2022 Jul 12;22(14):5206. doi: 10.3390/s22145206.
7
Kalman Filter-Based Fusion of LiDAR and Camera Data in Bird's Eye View for Multi-Object Tracking in Autonomous Vehicles.基于卡尔曼滤波器的激光雷达与相机数据融合在鸟瞰图中用于自动驾驶车辆的多目标跟踪
Sensors (Basel). 2024 Dec 3;24(23):7718. doi: 10.3390/s24237718.
8
Accurate 3-DoF Camera Geo-Localization via Ground-to-Satellite Image Matching.基于地-星图像匹配的精确 3-DoF 相机地理定位。
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):2682-2697. doi: 10.1109/TPAMI.2022.3189702. Epub 2023 Feb 3.
9
IRBEVF-Q: Optimization of Image-Radar Fusion Algorithm Based on Bird's Eye View Features.IRBEVF-Q:基于鸟瞰特征的图像-雷达融合算法优化
Sensors (Basel). 2024 Jul 16;24(14):4602. doi: 10.3390/s24144602.
10
A Self-Adaptive Feature Extraction Method for Aerial-view Geo-localization.一种用于鸟瞰地理定位的自适应特征提取方法。
IEEE Trans Image Process. 2024 Dec 12;PP. doi: 10.1109/TIP.2024.3513157.

引用本文的文献

1
A Method Combining Discrete Cosine Transform with Attention for Multi-Temporal Remote Sensing Image Matching.一种将离散余弦变换与注意力机制相结合的多时相遥感图像匹配方法。
Sensors (Basel). 2025 Feb 22;25(5):1345. doi: 10.3390/s25051345.

本文引用的文献

1
Vision-Centric BEV Perception: A Survey.以视觉为中心的鸟瞰图感知:综述
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):10978-10997. doi: 10.1109/TPAMI.2024.3449912. Epub 2024 Nov 6.
2
Convolutional Cross-View Pose Estimation.卷积跨视图姿态估计
IEEE Trans Pattern Anal Mach Intell. 2024 May;46(5):3813-3831. doi: 10.1109/TPAMI.2023.3346924. Epub 2024 Apr 3.