Suppr超能文献

Unified Depth-Guided Feature Fusion and Reranking for Hierarchical Place Recognition.

作者信息

Li Kunmo, Ou Yongsheng, Ning Jian, Kong Fanchang, Cai Haiyang, Li Haoyang

机构信息

School of Control Science and Engineering, Dalian University of Technology, Dalian 116024, China.

School of Computer Science, Wuhan University, Wuhan 430072, China.

出版信息

Sensors (Basel). 2025 Jun 29;25(13):4056. doi: 10.3390/s25134056.

Abstract

Visual Place Recognition (VPR) constitutes a pivotal task in the domains of computer vision and robotics. Prevailing VPR methods predominantly employ RGB-based features for query image retrieval and correspondence establishment. Nevertheless, such unimodal visual representations exhibit inherent susceptibility to environmental variations, inevitably degrading method precision. To address this problem, we propose a robust VPR framework integrating RGB and depth modalities. The architecture employs a coarse-to-fine paradigm, where global retrieval of top-N candidate images is performed using fused multimodal features, followed by a geometric verification of these candidates leveraging depth information. A Discrete Wavelet Transform Fusion (DWTF) module is proposed to generate robust multimodal global descriptors by effectively combining RGB and depth data using discrete wavelet transform. Furthermore, we introduce a Spiking Neuron Graph Matching (SNGM) module, which extracts geometric structure and spatial distance from depth data and employs graph matching for accurate depth feature correspondence. Extensive experiments on several VPR benchmarks demonstrate that our method achieves state-of-the-art performance while maintaining the best accuracy-efficiency trade-off.

摘要
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f63e/12252300/aeb42edcaa6c/sensors-25-04056-g001.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验