• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于基于视频的行人重识别的多级融合时空协同注意力

Multi-Level Fusion Temporal-Spatial Co-Attention for Video-Based Person Re-Identification.

作者信息

Pei Shengyu, Fan Xiaoping

机构信息

School of Automation, Central South University, Changsha 410075, China.

School of Information Technology and Management, Hunan University of Finance and Economics, Changsha 410205, China.

出版信息

Entropy (Basel). 2021 Dec 15;23(12):1686. doi: 10.3390/e23121686.

DOI:10.3390/e23121686
PMID:34945992
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8700156/
Abstract

A convolutional neural network can easily fall into local minima for insufficient data, and the needed training is unstable. Many current methods are used to solve these problems by adding pedestrian attributes, pedestrian postures, and other auxiliary information, but they require additional collection, which is time-consuming and laborious. Every video sequence frame has a different degree of similarity. In this paper, multi-level fusion temporal-spatial co-attention is adopted to improve person re-identification (reID). For a small dataset, the improved network can better prevent over-fitting and reduce the dataset limit. Specifically, the concept of knowledge evolution is introduced into video-based person re-identification to improve the backbone residual neural network (ResNet). The global branch, local branch, and attention branch are used in parallel for feature extraction. Three high-level features are embedded in the metric learning network to improve the network's generalization ability and the accuracy of video-based person re-identification. Simulation experiments are implemented on small datasets PRID2011 and iLIDS-VID, and the improved network can better prevent over-fitting. Experiments are also implemented on MARS and DukeMTMC-VideoReID, and the proposed method can be used to extract more feature information and improve the network's generalization ability. The results show that our method achieves better performance. The model achieves 90.15% Rank1 and 81.91% mAP on MARS.

摘要

卷积神经网络可能会因数据不足而轻易陷入局部最小值,且所需的训练不稳定。当前许多方法通过添加行人属性、行人姿态等辅助信息来解决这些问题,但这需要额外收集,既耗时又费力。每个视频序列帧都有不同程度的相似性。本文采用多级融合时空协同注意力来改进行人重识别(reID)。对于小数据集,改进后的网络能更好地防止过拟合并减少数据集限制。具体而言,将知识进化的概念引入基于视频的行人重识别,以改进骨干残差神经网络(ResNet)。全局分支、局部分支和注意力分支并行用于特征提取。将三个高级特征嵌入度量学习网络,以提高网络的泛化能力和基于视频的行人重识别的准确率。在小数据集PRID2011和iLIDS-VID上进行了仿真实验,改进后的网络能更好地防止过拟合。还在MARS和DukeMTMC-VideoReID上进行了实验,所提方法可用于提取更多特征信息并提高网络的泛化能力。结果表明,我们的方法取得了更好的性能。该模型在MARS上达到了90.15%的Rank1和81.91%的mAP。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/f5c995bc99a2/entropy-23-01686-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/2bd506bf77a5/entropy-23-01686-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/62505fdee097/entropy-23-01686-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/809ea784391c/entropy-23-01686-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/9a913b296f56/entropy-23-01686-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/217dffb936ed/entropy-23-01686-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/d7f2fb7341f6/entropy-23-01686-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/3b9f8d4aa1aa/entropy-23-01686-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/3b9d50c9c800/entropy-23-01686-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/c68ab3b9f840/entropy-23-01686-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/f5c995bc99a2/entropy-23-01686-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/2bd506bf77a5/entropy-23-01686-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/62505fdee097/entropy-23-01686-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/809ea784391c/entropy-23-01686-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/9a913b296f56/entropy-23-01686-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/217dffb936ed/entropy-23-01686-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/d7f2fb7341f6/entropy-23-01686-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/3b9f8d4aa1aa/entropy-23-01686-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/3b9d50c9c800/entropy-23-01686-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/c68ab3b9f840/entropy-23-01686-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/eeed/8700156/f5c995bc99a2/entropy-23-01686-g010.jpg

相似文献

1
Multi-Level Fusion Temporal-Spatial Co-Attention for Video-Based Person Re-Identification.用于基于视频的行人重识别的多级融合时空协同注意力
Entropy (Basel). 2021 Dec 15;23(12):1686. doi: 10.3390/e23121686.
2
Multi-scale Temporal Cues Learning for Video Person Re-Identification.用于视频人物重识别的多尺度时间线索学习
IEEE Trans Image Process. 2020 Feb 14. doi: 10.1109/TIP.2020.2972108.
3
Video-based person re-identification with complementary local and global features using a graph transformer.基于视频的人物再识别,使用图变换器融合互补的局部和全局特征。
Math Biosci Eng. 2024 Jul 23;21(7):6694-6709. doi: 10.3934/mbe.2024293.
4
Multi-granularity graph pooling for video-based person re-identification.基于视频的行人再识别的多粒度图池化。
Neural Netw. 2023 Mar;160:22-33. doi: 10.1016/j.neunet.2022.12.015. Epub 2022 Dec 28.
5
Video Person Re-identification by Temporal Residual Learning.基于时间残差学习的视频人物重识别
IEEE Trans Image Process. 2018 Oct 29. doi: 10.1109/TIP.2018.2878505.
6
Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature.基于端到端学习架构的混合深度表观-时间特征的视频人物再识别
Sensors (Basel). 2018 Oct 29;18(11):3669. doi: 10.3390/s18113669.
7
Video Person Re-Identification with Frame Sampling-Random Erasure and Mutual Information-Temporal Weight Aggregation.基于帧采样-随机擦除和互信息-时间权重聚合的视频行人再识别。
Sensors (Basel). 2022 Apr 15;22(8):3047. doi: 10.3390/s22083047.
8
Pedestrian re-identification based on attention mechanism and Multi-scale feature fusion.基于注意力机制和多尺度特征融合的行人再识别。
Math Biosci Eng. 2023 Aug 25;20(9):16913-16938. doi: 10.3934/mbe.2023754.
9
Adaptive Graph Representation Learning for Video Person Re-identification.用于视频人物重识别的自适应图表示学习
IEEE Trans Image Process. 2020 Jun 17;PP. doi: 10.1109/TIP.2020.3001693.
10
Exploring High-Order Spatio-Temporal Correlations From Skeleton for Person Re-Identification.从骨骼中探索用于行人重识别的高阶时空相关性。
IEEE Trans Image Process. 2023;32:949-963. doi: 10.1109/TIP.2023.3236144. Epub 2023 Jan 23.

本文引用的文献

1
A Two-Stream Dynamic Pyramid Representation Model for Video-Based Person Re-Identification.基于视频的人像再识别的双流动态金字塔表示模型。
IEEE Trans Image Process. 2021;30:6266-6276. doi: 10.1109/TIP.2021.3093759. Epub 2021 Jul 12.
2
Learning Generalisable Omni-Scale Representations for Person Re-Identification.学习用于行人重识别的通用全尺度表示。
IEEE Trans Pattern Anal Mach Intell. 2021 Mar 26;PP. doi: 10.1109/TPAMI.2021.3069237.
3
Identifying Visible Parts via Pose Estimation for Occluded Person Re-Identification.
基于姿态估计的遮挡行人再识别中可见部件的识别
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4624-4634. doi: 10.1109/TNNLS.2021.3059515. Epub 2022 Aug 31.
4
Adaptive Graph Representation Learning for Video Person Re-identification.用于视频人物重识别的自适应图表示学习
IEEE Trans Image Process. 2020 Jun 17;PP. doi: 10.1109/TIP.2020.3001693.
5
Ordered or Orderless: A Revisit for Video Based Person Re-Identification.有序或无序:基于视频的行人再识别研究回顾。
IEEE Trans Pattern Anal Mach Intell. 2021 Apr;43(4):1460-1466. doi: 10.1109/TPAMI.2020.2976969. Epub 2021 Mar 4.
6
Person Re-Identification With Deep Kronecker-Product Matching and Group-Shuffling Random Walk.基于深度 Kronecker 积匹配和分组洗牌随机游走的行人再识别。
IEEE Trans Pattern Anal Mach Intell. 2021 May;43(5):1649-1665. doi: 10.1109/TPAMI.2019.2954313. Epub 2021 Apr 1.
7
Person Re-Identification by Discriminative Selection in Video Ranking.基于视频排序的判别选择进行人像重识别
IEEE Trans Pattern Anal Mach Intell. 2016 Dec;38(12):2501-2514. doi: 10.1109/TPAMI.2016.2522418. Epub 2016 Jan 27.