• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于CLIP的行人重识别中领域差距的研究

An Investigation of the Domain Gap in CLIP-Based Person Re-Identification.

作者信息

Asperti Andrea, Naldi Leonardo, Fiorilla Salvatore

机构信息

Department of Informatics-Science and Engineering (DISI), University of Bologna, 40126 Bologna, Italy.

出版信息

Sensors (Basel). 2025 Jan 9;25(2):363. doi: 10.3390/s25020363.

DOI:10.3390/s25020363
PMID:39860732
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11769178/
Abstract

Person re-identification (re-id) is a critical computer vision task aimed at identifying individuals across multiple non-overlapping cameras, with wide-ranging applications in intelligent surveillance systems. Despite recent advances, the domain gap-performance degradation when models encounter unseen datasets-remains a critical challenge. CLIP-based models, leveraging multimodal pre-training, offer potential for mitigating this issue by aligning visual and textual representations. In this study, we provide a comprehensive quantitative analysis of the domain gap in CLIP-based re-id systems across standard benchmarks, including Market-1501, DukeMTMC-reID, MSMT17, and Airport, simulating real-world deployment conditions. We systematically measure the performance of these models in terms of mean average precision (mAP) and Rank-1 accuracy, offering insights into the challenges faced during dataset transitions. Our analysis highlights the specific advantages introduced by CLIP's visual-textual alignment and evaluates its contribution relative to strong image encoder baselines. Additionally, we evaluate the impact of extending training sets with non-domain-specific data and incorporating random erasing augmentation, achieving an average improvement of +4.3% in mAP and +4.0% in Rank-1 accuracy. Our findings underscore the importance of standardized benchmarks and systematic evaluations for enhancing reproducibility and guiding future research. This work contributes to a deeper understanding of the domain gap in re-id, while highlighting pathways for improving model robustness and generalization in diverse, real-world scenarios.

摘要

行人重识别(re-id)是一项关键的计算机视觉任务,旨在通过多个不重叠的摄像头识别个体,在智能监控系统中有广泛应用。尽管最近取得了进展,但当模型遇到未见数据集时的领域差距——性能下降——仍然是一个关键挑战。基于CLIP的模型利用多模态预训练,通过对齐视觉和文本表示,为缓解这一问题提供了潜力。在本研究中,我们对基于CLIP的re-id系统在包括Market-1501、DukeMTMC-reID、MSMT17和Airport等标准基准上的领域差距进行了全面的定量分析,模拟了实际部署条件。我们系统地根据平均精度均值(mAP)和Rank-1准确率来衡量这些模型的性能,深入了解数据集转换过程中面临的挑战。我们的分析突出了CLIP的视觉-文本对齐所带来的特定优势,并评估了其相对于强大的图像编码器基线的贡献。此外,我们评估了用非特定领域数据扩展训练集以及纳入随机擦除增强的影响,在mAP上平均提高了+4.3%,在Rank-1准确率上提高了+4.0%。我们的研究结果强调了标准化基准和系统评估对于提高可重复性和指导未来研究的重要性。这项工作有助于更深入地理解re-id中的领域差距,同时突出了在多样的现实场景中提高模型鲁棒性和泛化能力的途径。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/1981acbb889b/sensors-25-00363-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/3e75c2a12164/sensors-25-00363-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/c58ef881db8f/sensors-25-00363-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/2c0b6e038f49/sensors-25-00363-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/d40575bfada5/sensors-25-00363-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/9a84858e85d3/sensors-25-00363-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/6f1f0789d1a6/sensors-25-00363-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/89b9779e5a83/sensors-25-00363-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/1981acbb889b/sensors-25-00363-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/3e75c2a12164/sensors-25-00363-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/c58ef881db8f/sensors-25-00363-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/2c0b6e038f49/sensors-25-00363-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/d40575bfada5/sensors-25-00363-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/9a84858e85d3/sensors-25-00363-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/6f1f0789d1a6/sensors-25-00363-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/89b9779e5a83/sensors-25-00363-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bfa1/11769178/1981acbb889b/sensors-25-00363-g008.jpg

相似文献

1
An Investigation of the Domain Gap in CLIP-Based Person Re-Identification.基于CLIP的行人重识别中领域差距的研究
Sensors (Basel). 2025 Jan 9;25(2):363. doi: 10.3390/s25020363.
2
A Multi-Attention Approach for Person Re-Identification Using Deep Learning.基于深度学习的多注意力机制行人再识别方法。
Sensors (Basel). 2023 Apr 2;23(7):3678. doi: 10.3390/s23073678.
3
Synth-CLIP: Synthetic data make CLIP generalize better in data-limited scenarios.合成CLIP:合成数据使CLIP在数据有限的场景中泛化能力更强。
Neural Netw. 2025 Apr;184:107083. doi: 10.1016/j.neunet.2024.107083. Epub 2024 Dec 30.
4
Multi-Domain Adversarial Feature Generalization for Person Re-Identification.多领域对抗特征泛化的行人再识别
IEEE Trans Image Process. 2021;30:1596-1607. doi: 10.1109/TIP.2020.3046864. Epub 2021 Jan 11.
5
Deep Learning for Person Re-Identification: A Survey and Outlook.用于行人重识别的深度学习:综述与展望
IEEE Trans Pattern Anal Mach Intell. 2022 Jun;44(6):2872-2893. doi: 10.1109/TPAMI.2021.3054775. Epub 2022 May 5.
6
Approaches to Improve the Quality of Person Re-Identification for Practical Use.提高实际应用中人物重新识别质量的方法。
Sensors (Basel). 2023 Aug 24;23(17):7382. doi: 10.3390/s23177382.
7
ES-Net: Erasing Salient Parts to Learn More in Re-Identification.ES-Net:擦除显著部分以在再识别中学习更多。
IEEE Trans Image Process. 2021;30:1676-1686. doi: 10.1109/TIP.2020.3046904. Epub 2021 Jan 11.
8
Knowledge-Preserving continual person re-identification using Graph Attention Network.使用图注意力网络的知识保留持续人物再识别
Neural Netw. 2023 Apr;161:105-115. doi: 10.1016/j.neunet.2023.01.033. Epub 2023 Feb 1.
9
Unsupervised Cross Domain Person Re-Identification by Multi-Loss Optimization Learning.无监督跨域行人再识别的多损失优化学习。
IEEE Trans Image Process. 2021;30:2935-2946. doi: 10.1109/TIP.2021.3056889. Epub 2021 Feb 12.
10
Enhancing MRI brain tumor classification: A comprehensive approach integrating real-life scenario simulation and augmentation techniques.增强 MRI 脑肿瘤分类:一种综合方法,集成真实场景模拟和增强技术。
Phys Med. 2024 Nov;127:104841. doi: 10.1016/j.ejmp.2024.104841. Epub 2024 Nov 2.

引用本文的文献

1
Cross-Modal Weakly Supervised RGB-D Salient Object Detection with a Focus on Filamentary Structures.关注丝状结构的跨模态弱监督RGB-D显著目标检测
Sensors (Basel). 2025 May 9;25(10):2990. doi: 10.3390/s25102990.

本文引用的文献

1
A Generative Approach to Person Reidentification.一种用于行人重识别的生成方法。
Sensors (Basel). 2024 Feb 15;24(4):1240. doi: 10.3390/s24041240.
2
Discriminatively Unsupervised Learning Person Re-Identification via Considering Complicated Images.基于复杂图像的判别式无监督行人再识别学习。
Sensors (Basel). 2023 Mar 20;23(6):3259. doi: 10.3390/s23063259.
3
Deep Learning for Person Re-Identification: A Survey and Outlook.用于行人重识别的深度学习:综述与展望
IEEE Trans Pattern Anal Mach Intell. 2022 Jun;44(6):2872-2893. doi: 10.1109/TPAMI.2021.3054775. Epub 2022 May 5.
4
Deep Coattention-Based Comparator for Relative Representation Learning in Person Re-Identification.基于深度协同注意力的相对表示学习在行人再识别中的应用。
IEEE Trans Neural Netw Learn Syst. 2021 Feb;32(2):722-735. doi: 10.1109/TNNLS.2020.2979190. Epub 2021 Feb 4.
5
A Systematic Evaluation and Benchmark for Person Re-Identification: Features, Metrics, and Datasets.人物再识别的系统评估与基准:特征、指标和数据集
IEEE Trans Pattern Anal Mach Intell. 2019 Mar;41(3):523-536. doi: 10.1109/TPAMI.2018.2807450. Epub 2018 Feb 19.
6
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
7
Fast Feature Pyramids for Object Detection.快速目标检测特征金字塔。
IEEE Trans Pattern Anal Mach Intell. 2014 Aug;36(8):1532-45. doi: 10.1109/TPAMI.2014.2300479.
8
Object detection with discriminatively trained part-based models.基于判别式训练的部件模型的目标检测。
IEEE Trans Pattern Anal Mach Intell. 2010 Sep;32(9):1627-45. doi: 10.1109/TPAMI.2009.167.
9
Faster and better: a machine learning approach to corner detection.更快更好:一种用于角点检测的机器学习方法。
IEEE Trans Pattern Anal Mach Intell. 2010 Jan;32(1):105-19. doi: 10.1109/TPAMI.2008.275.