• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

跨视图地理定位的联合表示学习和关键点检测。

Joint Representation Learning and Keypoint Detection for Cross-View Geo-Localization.

出版信息

IEEE Trans Image Process. 2022;31:3780-3792. doi: 10.1109/TIP.2022.3175601. Epub 2022 Jun 2.

DOI:10.1109/TIP.2022.3175601
PMID:35604972
Abstract

In this paper, we study the cross-view geo-localization problem to match images from different viewpoints. The key motivation underpinning this task is to learn a discriminative viewpoint-invariant visual representation. Inspired by the human visual system for mining local patterns, we propose a new framework called RK-Net to jointly learn the discriminative Representation and detect salient Keypoints with a single Network. Specifically, we introduce a Unit Subtraction Attention Module (USAM) that can automatically discover representative keypoints from feature maps and draw attention to the salient regions. USAM contains very few learning parameters but yields significant performance improvement and can be easily plugged into different networks. We demonstrate through extensive experiments that (1) by incorporating USAM, RK-Net facilitates end-to-end joint learning without the prerequisite of extra annotations. Representation learning and keypoint detection are two highly-related tasks. Representation learning aids keypoint detection. Keypoint detection, in turn, enriches the model capability against large appearance changes caused by viewpoint variants. (2) USAM is easy to implement and can be integrated with existing methods, further improving the state-of-the-art performance. We achieve competitive geo-localization accuracy on three challenging datasets, i. e., University-1652, CVUSA and CVACT. Our code is available at https://github.com/AggMan96/RK-Net.

摘要

在本文中,我们研究了跨视图地理定位问题,以匹配来自不同视点的图像。这项任务的关键动机是学习具有辨别力的不变视点视觉表示。受人类视觉系统挖掘局部模式的启发,我们提出了一个名为 RK-Net 的新框架,该框架使用单个网络联合学习有辨别力的表示和检测显著的关键点。具体来说,我们引入了一个单元减法注意力模块(USAM),它可以自动从特征图中发现有代表性的关键点,并将注意力吸引到显著区域。USAM 包含很少的学习参数,但能显著提高性能,并且可以轻松地插入到不同的网络中。我们通过广泛的实验证明:(1)通过引入 USAM,RK-Net 促进了端到端的联合学习,而无需额外的注释。表示学习和关键点检测是两个高度相关的任务。表示学习有助于关键点检测。反过来,关键点检测丰富了模型的能力,以应对由视点变体引起的大的外观变化。(2)USAM 易于实现,并且可以与现有的方法集成,进一步提高了最先进的性能。我们在三个具有挑战性的数据集上实现了有竞争力的地理定位精度,即 University-1652、CVUSA 和 CVACT。我们的代码可在 https://github.com/AggMan96/RK-Net 获得。

相似文献

1
Joint Representation Learning and Keypoint Detection for Cross-View Geo-Localization.跨视图地理定位的联合表示学习和关键点检测。
IEEE Trans Image Process. 2022;31:3780-3792. doi: 10.1109/TIP.2022.3175601. Epub 2022 Jun 2.
2
Perspectively Equivariant Keypoint Learning for Omnidirectional Images.全方位图像的透视等变关键点学习。
IEEE Trans Image Process. 2023;32:2552-2567. doi: 10.1109/TIP.2023.3270032. Epub 2023 May 5.
3
UAV's Status Is Worth Considering: A Fusion Representations Matching Method for Geo-Localization.无人机的状态值得考虑:一种用于地理定位的融合表示匹配方法。
Sensors (Basel). 2023 Jan 8;23(2):720. doi: 10.3390/s23020720.
4
Variational Structured Attention Networks for Deep Visual Representation Learning.用于深度视觉表征学习的变分结构化注意力网络
IEEE Trans Image Process. 2022 Mar 2;PP. doi: 10.1109/TIP.2021.3137647.
5
Dynamic Keypoint Detection Network for Image Matching.用于图像匹配的动态关键点检测网络
IEEE Trans Pattern Anal Mach Intell. 2023 Dec;45(12):14404-14419. doi: 10.1109/TPAMI.2023.3307889. Epub 2023 Nov 3.
6
Object and spatial discrimination makes weakly supervised local feature better.目标和空间辨别能力使弱监督局部特征更优。
Neural Netw. 2024 Dec;180:106697. doi: 10.1016/j.neunet.2024.106697. Epub 2024 Sep 12.
7
Tasks Integrated Networks: Joint Detection and Retrieval for Image Search.任务集成网络:图像搜索的联合检测与检索。
IEEE Trans Pattern Anal Mach Intell. 2022 Jan;44(1):456-473. doi: 10.1109/TPAMI.2020.3009758. Epub 2021 Dec 7.
8
Attention-Guided Discriminative Region Localization and Label Distribution Learning for Bone Age Assessment.注意力引导的判别性区域定位和标签分布学习在骨龄评估中的应用。
IEEE J Biomed Health Inform. 2022 Mar;26(3):1208-1218. doi: 10.1109/JBHI.2021.3095128. Epub 2022 Mar 7.
9
Learning Semantic-Aware Local Features for Long Term Visual Localization.学习用于长期视觉定位的语义感知局部特征。
IEEE Trans Image Process. 2022;31:4842-4855. doi: 10.1109/TIP.2022.3187565. Epub 2022 Jul 20.
10
Multi-Task Structure-Aware Context Modeling for Robust Keypoint-Based Object Tracking.用于基于关键点的稳健目标跟踪的多任务结构感知上下文建模
IEEE Trans Pattern Anal Mach Intell. 2019 Apr;41(4):915-927. doi: 10.1109/TPAMI.2018.2818132. Epub 2018 Mar 22.

引用本文的文献

1
UPANets: Learning from the Universal Pixel Attention Neworks.UPANets:从通用像素注意力网络中学习。
Entropy (Basel). 2022 Sep 4;24(9):1243. doi: 10.3390/e24091243.