• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

学习用于长期视觉定位的语义感知局部特征。

Learning Semantic-Aware Local Features for Long Term Visual Localization.

出版信息

IEEE Trans Image Process. 2022;31:4842-4855. doi: 10.1109/TIP.2022.3187565. Epub 2022 Jul 20.

DOI:10.1109/TIP.2022.3187565
PMID:35830407
Abstract

Extracting robust and discriminative local features from images plays a vital role for long term visual localization, whose challenges are mainly caused by the severe appearance differences between matching images due to the day-night illuminations, seasonal changes, and human activities. Existing solutions resort to jointly learning both keypoints and their descriptors in an end-to-end manner, leveraged on large number of annotations of point correspondence which are harvested from the structure from motion and depth estimation algorithms. While these methods show improved performance over non-deep methods or those two-stage deep methods, i.e., detection and then description, they are still struggled to conquer the problems encountered in long term visual localization. Since the intrinsic semantics are invariant to the local appearance changes, this paper proposes to learn semantic-aware local features in order to improve robustness of local feature matching for long term localization. Based on a state of the art CNN architecture for local feature learning, i.e., ASLFeat, this paper leverages on the semantic information from an off-the-shelf semantic segmentation network to learn semantic-aware feature maps. The learned correspondence-aware feature descriptors and semantic features are then merged to form the final feature descriptors, for which the improved feature matching ability has been observed in experiments. In addition, the learned semantics embedded in the features can be further used to filter out noisy keypoints, leading to additional accuracy improvement and faster matching speed. Experiments on two popular long term visual localization benchmarks (Aachen Day and Night v1.1, Robotcar Seasons) and one challenging indoor benchmark (InLoc) demonstrate encouraging improvements of the localization accuracy over its counterpart and other competitive methods.

摘要

从图像中提取稳健且具有鉴别力的局部特征对于长期视觉定位至关重要,其挑战主要源于匹配图像之间由于日夜光照、季节变化和人类活动等因素导致的严重外观差异。现有的解决方案倾向于以端到端的方式联合学习关键点及其描述符,利用从运动结构和深度估计算法中提取的大量点对应标注。虽然这些方法在非深度方法或两阶段深度方法(即检测然后描述)方面表现出了改进的性能,但它们仍然难以克服长期视觉定位中遇到的问题。由于内在语义对于局部外观变化是不变的,因此本文提出了学习语义感知的局部特征,以提高局部特征匹配的鲁棒性,从而实现长期本地化。基于一种用于局部特征学习的先进卷积神经网络架构(ASLFeat),本文利用来自现成语义分割网络的语义信息来学习语义感知特征图。然后将学习到的对应感知特征描述符和语义特征合并形成最终的特征描述符,实验中观察到改进的特征匹配能力。此外,嵌入在特征中的学习语义可进一步用于过滤掉噪声关键点,从而实现额外的准确性提高和更快的匹配速度。在两个流行的长期视觉定位基准(Aachen Day and Night v1.1、Robotcar Seasons)和一个具有挑战性的室内基准(InLoc)上进行的实验表明,与基准和其他竞争方法相比,定位精度有了令人鼓舞的提高。

相似文献

1
Learning Semantic-Aware Local Features for Long Term Visual Localization.学习用于长期视觉定位的语义感知局部特征。
IEEE Trans Image Process. 2022;31:4842-4855. doi: 10.1109/TIP.2022.3187565. Epub 2022 Jul 20.
2
SFA-MDEN: Semantic-Feature-Aided Monocular Depth Estimation Network Using Dual Branches.SFA-MDEN:基于语义特征辅助的双通道单目深度估计网络。
Sensors (Basel). 2021 Aug 13;21(16):5476. doi: 10.3390/s21165476.
3
Object and spatial discrimination makes weakly supervised local feature better.目标和空间辨别能力使弱监督局部特征更优。
Neural Netw. 2024 Dec;180:106697. doi: 10.1016/j.neunet.2024.106697. Epub 2024 Sep 12.
4
Scene-graph-driven semantic feature matching for monocular digestive endoscopy.基于场景图的语义特征匹配用于单目消化内镜
Comput Biol Med. 2022 Jul;146:105616. doi: 10.1016/j.compbiomed.2022.105616. Epub 2022 May 16.
5
SVS-VPR: A Semantic Visual and Spatial Information-Based Hierarchical Visual Place Recognition for Autonomous Navigation in Challenging Environmental Conditions.SVS-VPR:一种基于语义视觉和空间信息的分层视觉场所识别方法,用于在具有挑战性的环境条件下进行自主导航。
Sensors (Basel). 2024 Jan 30;24(3):906. doi: 10.3390/s24030906.
6
An Appearance-Semantic Descriptor with Coarse-to-Fine Matching for Robust VPR.一种具有从粗到细匹配的外观语义描述符用于鲁棒视觉位置识别
Sensors (Basel). 2024 Mar 29;24(7):2203. doi: 10.3390/s24072203.
7
LIO-CSI: LiDAR inertial odometry with loop closure combined with semantic information.LIO-CSI:结合语义信息的具有闭环检测功能的激光雷达惯性里程计。
PLoS One. 2021 Dec 8;16(12):e0261053. doi: 10.1371/journal.pone.0261053. eCollection 2021.
8
Robust facial landmark detection by cross-order cross-semantic deep network.基于跨阶跨语义深度网络的鲁棒人脸关键点检测
Neural Netw. 2021 Apr;136:233-243. doi: 10.1016/j.neunet.2020.11.001. Epub 2020 Nov 16.
9
TS-CAM: Token Semantic Coupled Attention Map for Weakly Supervised Object Localization.TS-CAM:用于弱监督目标定位的令牌语义耦合注意力图
IEEE Trans Neural Netw Learn Syst. 2024 Jul;35(7):9109-9121. doi: 10.1109/TNNLS.2022.3218471. Epub 2024 Jul 8.
10
Joint Representation Learning and Keypoint Detection for Cross-View Geo-Localization.跨视图地理定位的联合表示学习和关键点检测。
IEEE Trans Image Process. 2022;31:3780-3792. doi: 10.1109/TIP.2022.3175601. Epub 2022 Jun 2.

引用本文的文献

1
Multiscale Cascaded Attention Network for Saliency Detection Based on ResNet.基于 ResNet 的多尺度级联注意力网络的显著目标检测
Sensors (Basel). 2022 Dec 16;22(24):9950. doi: 10.3390/s22249950.