• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于帧采样-随机擦除和互信息-时间权重聚合的视频行人再识别。

Video Person Re-Identification with Frame Sampling-Random Erasure and Mutual Information-Temporal Weight Aggregation.

机构信息

Information and Communication Engineering, Electronics Information Engineering College, Changchun University of Science and Technology, Changchun 130022, China.

High-Speed Railway Comprehensive Technical College, Jilin Railway Technology College, Jilin 132299, China.

出版信息

Sensors (Basel). 2022 Apr 15;22(8):3047. doi: 10.3390/s22083047.

DOI:10.3390/s22083047
PMID:35459030
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9032512/
Abstract

Partial occlusion and background clutter in camera video surveillance affect the accuracy of video-based person re-identification (re-ID). To address these problems, we propose a person re-ID method based on random erasure of frame sampling and temporal weight aggregation of mutual information of partial and global features. First, for the case in which the target person is interfered or partially occluded, the frame sampling-random erasure (FSE) method is used for data enhancement to effectively alleviate the occlusion problem, improve the generalization ability of the model, and match persons more accurately. Second, to further improve the re-ID accuracy of video-based persons and learn more discriminative feature representations, we use a ResNet-50 network to extract global and partial features and fuse these features to obtain frame-level features. In the time dimension, based on a mutual information-temporal weight aggregation (MI-TWA) module, the partial features are added according to different weights and the global features are added according to equal weights and connected to output sequence features. The proposed method is extensively experimented on three public video datasets, MARS, DukeMTMC-VideoReID, and PRID-2011; the mean average precision (mAP) values are 82.4%, 94.1%, and 95.3% and Rank-1 values are 86.4%, 94.8%, and 95.2%, respectively.

摘要

相机视频监控中的部分遮挡和背景杂波会影响基于视频的人员重新识别(re-ID)的准确性。针对这些问题,我们提出了一种基于帧采样随机擦除和部分与全局特征互信息时间权重聚合的人员重新识别方法。首先,对于目标人员受到干扰或部分遮挡的情况,使用帧采样随机擦除(FSE)方法进行数据增强,有效缓解遮挡问题,提高模型的泛化能力,更准确地匹配人员。其次,为了进一步提高基于视频的人员重新识别的准确性,并学习更具判别力的特征表示,我们使用 ResNet-50 网络提取全局和部分特征,并融合这些特征以获得帧级特征。在时间维度上,基于互信息-时间权重聚合(MI-TWA)模块,根据不同的权重添加部分特征,并根据等权重添加全局特征,并连接输出序列特征。在三个公共视频数据集 MARS、DukeMTMC-VideoReID 和 PRID-2011 上进行了广泛的实验,平均精度(mAP)值分别为 82.4%、94.1%和 95.3%,排名第一(Rank-1)值分别为 86.4%、94.8%和 95.2%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/08a5/9032512/35b3301bf32b/sensors-22-03047-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/08a5/9032512/7ae1b4679270/sensors-22-03047-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/08a5/9032512/7b6566e7b0d8/sensors-22-03047-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/08a5/9032512/35b3301bf32b/sensors-22-03047-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/08a5/9032512/7ae1b4679270/sensors-22-03047-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/08a5/9032512/7b6566e7b0d8/sensors-22-03047-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/08a5/9032512/35b3301bf32b/sensors-22-03047-g003.jpg

相似文献

1
Video Person Re-Identification with Frame Sampling-Random Erasure and Mutual Information-Temporal Weight Aggregation.基于帧采样-随机擦除和互信息-时间权重聚合的视频行人再识别。
Sensors (Basel). 2022 Apr 15;22(8):3047. doi: 10.3390/s22083047.
2
Video-based person re-identification with complementary local and global features using a graph transformer.基于视频的人物再识别,使用图变换器融合互补的局部和全局特征。
Math Biosci Eng. 2024 Jul 23;21(7):6694-6709. doi: 10.3934/mbe.2024293.
3
Multi-Level Fusion Temporal-Spatial Co-Attention for Video-Based Person Re-Identification.用于基于视频的行人重识别的多级融合时空协同注意力
Entropy (Basel). 2021 Dec 15;23(12):1686. doi: 10.3390/e23121686.
4
Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature.基于端到端学习架构的混合深度表观-时间特征的视频人物再识别
Sensors (Basel). 2018 Oct 29;18(11):3669. doi: 10.3390/s18113669.
5
Anchor Association Learning for Unsupervised Video Person Re-Identification.用于无监督视频人物重识别的锚点关联学习
IEEE Trans Neural Netw Learn Syst. 2024 Jan;35(1):1013-1024. doi: 10.1109/TNNLS.2022.3179133. Epub 2024 Jan 4.
6
Iterative Local-Global Collaboration Learning towards One-Shot Video Person Re-Identification.面向一次性视频人物重识别的迭代局部-全局协作学习
IEEE Trans Image Process. 2020 Oct 2;PP. doi: 10.1109/TIP.2020.3026625.
7
Motion Feature Aggregation for Video-Based Person Re-Identification.基于视频的行人重识别中的运动特征聚合
IEEE Trans Image Process. 2022;31:3908-3919. doi: 10.1109/TIP.2022.3175593. Epub 2022 Jun 9.
8
3-D PersonVLAD: Learning Deep Global Representations for Video-Based Person Reidentification.3-D 人 VLAD:基于视频的人体再识别的深度学习全局表示方法。
IEEE Trans Neural Netw Learn Syst. 2019 Nov;30(11):3347-3359. doi: 10.1109/TNNLS.2019.2891244. Epub 2019 Feb 1.
9
Adaptive Graph Representation Learning for Video Person Re-identification.用于视频人物重识别的自适应图表示学习
IEEE Trans Image Process. 2020 Jun 17;PP. doi: 10.1109/TIP.2020.3001693.
10
A Two-Stream Dynamic Pyramid Representation Model for Video-Based Person Re-Identification.基于视频的人像再识别的双流动态金字塔表示模型。
IEEE Trans Image Process. 2021;30:6266-6276. doi: 10.1109/TIP.2021.3093759. Epub 2021 Jul 12.

引用本文的文献

1
Person Re-Identification with Improved Performance by Incorporating Focal Tversky Loss in AGW Baseline.基于 AGW 基线融合焦点 Tversky 损失改进性能的行人再识别。
Sensors (Basel). 2022 Dec 15;22(24):9852. doi: 10.3390/s22249852.

本文引用的文献

1
Video Person Re-identification by Temporal Residual Learning.基于时间残差学习的视频人物重识别
IEEE Trans Image Process. 2018 Oct 29. doi: 10.1109/TIP.2018.2878505.