• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于区域增强和跨模态注意力的可见-红外行人重识别

Visible-infrared person re-identification with region-based augmentation and cross modality attention.

作者信息

Guo Yuwei, Zhang Wenhao, Jiao Licheng, Wang Shuang, Wang Shuo, Liu Fang

机构信息

Key Laboratory of Intelligent Perception and Image Understanding of the Ministry of Education of China, School of Artificial Intelligence, International Research Center of Intelligent Perception and Computation, Xidian University, Xi'an, 710071, China.

School of Computer Science, The University of Birmingham, Birmingham, B15 2TT, UK.

出版信息

Sci Rep. 2025 May 25;15(1):18225. doi: 10.1038/s41598-025-01979-z.

DOI:10.1038/s41598-025-01979-z
PMID:40414905
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12104378/
Abstract

Visible-infrared person re-identification (VI-ReID) aims to search the same pedestrian of interest across visible and infrared modalities. Existing models mainly focus on compensating for modality-specific information to reduce modality variation. However, these methods often introduce interfering information and lead to higher computational overhead when generating the corresponding images or features. Additionally, the pedestrian region characteristics in VI-ReID are not effectively utilized, thus resulting in ambiguous or unnatural images. To address these issues, it is critical to leverage pedestrian attentive features and learn modality-complete and -consistent representation. In this paper, a novel Region-based Augmentation and Cross Modality Attention (RACA) model is proposed, focusing on the pedestrian regions to efficiently compensate for missing modality-specific features. Specifically, we propose a region-based data augmentation module PedMix to enhance pedestrian region coherence by mixing the corresponding regions from different modalities, thus generating more natural images. Moreover, a lightweight hybrid compensation module, i.e., a Modality Feature Transfer (MFT) module, is proposed to integrate cross attention and convolution networks to avoid introducing interfering information while preserving minimal computational overhead. Extensive experiments conducted on the benchmark SYSU-MM01 and RegDB datasets demonstrated the effectiveness of our proposed RACA model.

摘要

可见-红外行人重识别(VI-ReID)旨在跨可见光和红外模态搜索同一感兴趣的行人。现有模型主要专注于补偿特定模态信息以减少模态差异。然而,这些方法在生成相应图像或特征时常常引入干扰信息并导致更高的计算开销。此外,VI-ReID中的行人区域特征未得到有效利用,从而导致图像模糊或不自然。为了解决这些问题,利用行人注意力特征并学习模态完整且一致的表示至关重要。本文提出了一种新颖的基于区域的增强与跨模态注意力(RACA)模型,聚焦于行人区域以有效补偿缺失的特定模态特征。具体而言,我们提出了一个基于区域的数据增强模块PedMix,通过混合来自不同模态的相应区域来增强行人区域的连贯性,从而生成更自然的图像。此外,还提出了一个轻量级混合补偿模块,即模态特征转移(MFT)模块,以整合交叉注意力和卷积网络,在保持最小计算开销的同时避免引入干扰信息。在基准SYSU-MM01和RegDB数据集上进行的大量实验证明了我们提出的RACA模型的有效性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/e0ab9e54e572/41598_2025_1979_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/0d5c94187b1c/41598_2025_1979_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/8618a62100dd/41598_2025_1979_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/3a1c448c1b1c/41598_2025_1979_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/8ea344f28286/41598_2025_1979_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/2ab6979644ad/41598_2025_1979_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/8e3de9d61e64/41598_2025_1979_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/2b196a8d412c/41598_2025_1979_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/e0ab9e54e572/41598_2025_1979_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/0d5c94187b1c/41598_2025_1979_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/8618a62100dd/41598_2025_1979_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/3a1c448c1b1c/41598_2025_1979_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/8ea344f28286/41598_2025_1979_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/2ab6979644ad/41598_2025_1979_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/8e3de9d61e64/41598_2025_1979_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/2b196a8d412c/41598_2025_1979_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1a67/12104378/e0ab9e54e572/41598_2025_1979_Fig8_HTML.jpg

相似文献

1
Visible-infrared person re-identification with region-based augmentation and cross modality attention.基于区域增强和跨模态注意力的可见-红外行人重识别
Sci Rep. 2025 May 25;15(1):18225. doi: 10.1038/s41598-025-01979-z.
2
Flexible Body Partition-Based Adversarial Learning for Visible Infrared Person Re-Identification.基于柔性体分区的可见光红外行人再识别对抗学习
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4676-4687. doi: 10.1109/TNNLS.2021.3059713. Epub 2022 Aug 31.
3
Nystromformer based cross-modality transformer for visible-infrared person re-identification.基于Nystromformer的跨模态变压器用于可见光-红外行人重识别
Sci Rep. 2025 May 9;15(1):16224. doi: 10.1038/s41598-025-01226-5.
4
Cross-Modality Person Re-Identification Method with Joint-Modality Generation and Feature Enhancement.基于联合模态生成与特征增强的跨模态行人重识别方法
Entropy (Basel). 2024 Aug 13;26(8):681. doi: 10.3390/e26080681.
5
BiFFN: Bi-Frequency Guided Feature Fusion Network for Visible-Infrared Person Re-Identification.BiFFN:用于可见光-红外人体重识别的双频引导特征融合网络。
Sensors (Basel). 2025 Feb 20;25(5):1298. doi: 10.3390/s25051298.
6
CycleTrans: Learning Neutral Yet Discriminative Features via Cycle Construction for Visible- Infrared Person Re-Identification.循环变换:通过循环构建学习用于可见光-红外行人重识别的中性且有区分性的特征
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):5469-5479. doi: 10.1109/TNNLS.2024.3382937. Epub 2025 Feb 28.
7
Cross-Modality Person Re-Identification via Local Paired Graph Attention Network.基于局部成对图注意网络的跨模态人像再识别。
Sensors (Basel). 2023 Apr 15;23(8):4011. doi: 10.3390/s23084011.
8
Visible-Infrared Person Re-Identification via Partially Interactive Collaboration.基于部分交互协作的可见光-近红外行人再识别。
IEEE Trans Image Process. 2022;31:6951-6963. doi: 10.1109/TIP.2022.3217697. Epub 2022 Nov 8.
9
Translation, Association and Augmentation: Learning Cross-Modality Re-Identification From Single-Modality Annotation.翻译、关联与增强:从单模态标注中学习跨模态重新识别
IEEE Trans Image Process. 2023;32:5099-5113. doi: 10.1109/TIP.2023.3310338. Epub 2023 Sep 12.
10
Visible-Infrared Person Re-Identification With Modality-Specific Memory Network.基于模态特定记忆网络的可见-近红外人像再识别
IEEE Trans Image Process. 2022;31:7165-7178. doi: 10.1109/TIP.2022.3220408. Epub 2022 Nov 16.

本文引用的文献

1
Visible-Infrared Person Re-Identification With Modality-Specific Memory Network.基于模态特定记忆网络的可见-近红外人像再识别
IEEE Trans Image Process. 2022;31:7165-7178. doi: 10.1109/TIP.2022.3220408. Epub 2022 Nov 16.
2
Visible-Infrared Person Re-Identification via Partially Interactive Collaboration.基于部分交互协作的可见光-近红外行人再识别。
IEEE Trans Image Process. 2022;31:6951-6963. doi: 10.1109/TIP.2022.3217697. Epub 2022 Nov 8.
3
RBDF: Reciprocal Bidirectional Framework for Visible Infrared Person Reidentification.
RBDF:可见光红外行人再识别的双向互反框架。
IEEE Trans Cybern. 2022 Oct;52(10):10988-10998. doi: 10.1109/TCYB.2022.3183395. Epub 2022 Sep 19.
4
Structure-Aware Positional Transformer for Visible-Infrared Person Re-Identification.基于结构感知的可见光-红外跨模态行人重识别的位置变换模型
IEEE Trans Image Process. 2022;31:2352-2364. doi: 10.1109/TIP.2022.3141868. Epub 2022 Mar 15.
5
SFANet: A Spectrum-Aware Feature Augmentation Network for Visible-Infrared Person Reidentification.SFANet:一种用于可见光-红外行人重识别的频谱感知特征增强网络。
IEEE Trans Neural Netw Learn Syst. 2023 Apr;34(4):1958-1971. doi: 10.1109/TNNLS.2021.3105702. Epub 2023 Apr 4.
6
Flexible Body Partition-Based Adversarial Learning for Visible Infrared Person Re-Identification.基于柔性体分区的可见光红外行人再识别对抗学习
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4676-4687. doi: 10.1109/TNNLS.2021.3059713. Epub 2022 Aug 31.
7
Deep Learning for Person Re-Identification: A Survey and Outlook.用于行人重识别的深度学习:综述与展望
IEEE Trans Pattern Anal Mach Intell. 2022 Jun;44(6):2872-2893. doi: 10.1109/TPAMI.2021.3054775. Epub 2022 May 5.
8
Squeeze-and-Excitation Networks.挤压激励网络。
IEEE Trans Pattern Anal Mach Intell. 2020 Aug;42(8):2011-2023. doi: 10.1109/TPAMI.2019.2913372. Epub 2019 Apr 29.
9
Person Recognition System Based on a Combination of Body Images from Visible Light and Thermal Cameras.基于可见光和热成像相机人体图像组合的人员识别系统
Sensors (Basel). 2017 Mar 16;17(3):605. doi: 10.3390/s17030605.