• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

IAUnet:用于人体重识别的全局上下文感知特征学习。

IAUnet: Global Context-Aware Feature Learning for Person Reidentification.

出版信息

IEEE Trans Neural Netw Learn Syst. 2021 Oct;32(10):4460-4474. doi: 10.1109/TNNLS.2020.3017939. Epub 2021 Oct 5.

DOI:10.1109/TNNLS.2020.3017939
PMID:32877342
Abstract

Person reidentification (reID) by convolutional neural network (CNN)-based networks has achieved favorable performance in recent years. However, most of existing CNN-based methods do not take full advantage of spatial-temporal context modeling. In fact, the global spatial-temporal context can greatly clarify local distractions to enhance the target feature representation. To comprehensively leverage the spatial-temporal context information, in this work, we present a novel block, interaction-aggregation-update (IAU), for high-performance person reID. First, the spatial-temporal IAU (STIAU) module is introduced. STIAU jointly incorporates two types of contextual interactions into a CNN framework for target feature learning. Here, the spatial interactions learn to compute the contextual dependencies between different body parts of a single frame, while the temporal interactions are used to capture the contextual dependencies between the same body parts across all frames. Furthermore, a channel IAU (CIAU) module is designed to model the semantic contextual interactions between channel features to enhance the feature representation, especially for small-scale visual cues and body parts. Therefore, the IAU block enables the feature to incorporate the globally spatial, temporal, and channel context. It is lightweight, end-to-end trainable, and can be easily plugged into existing CNNs to form IAUnet. The experiments show that IAUnet performs favorably against state of the art on both image and video reID tasks and achieves compelling results on a general object categorization task. The source code is available at https://github.com/blue-blue272/ImgReID-IAnet.

摘要

基于卷积神经网络 (CNN) 的人物重识别 (reID) 在近年来取得了良好的性能。然而,现有的大多数基于 CNN 的方法并没有充分利用时空上下文建模。事实上,全局时空上下文可以极大地澄清局部干扰,从而增强目标特征表示。为了全面利用时空上下文信息,在这项工作中,我们提出了一种新颖的块,即交互聚合更新 (IAU),用于高性能的人物 reID。首先,引入了时空 IAU (STIAU) 模块。STIAU 将两种类型的上下文交互联合纳入 CNN 框架中,以学习目标特征。这里,空间交互用于学习计算单个帧中不同身体部位之间的上下文依赖关系,而时间交互用于捕获所有帧中相同身体部位之间的上下文依赖关系。此外,设计了一个通道 IAU (CIAU) 模块来建模通道特征之间的语义上下文交互,以增强特征表示,特别是对于小尺度的视觉线索和身体部位。因此,IAU 块使特征能够融合全局空间、时间和通道上下文。它轻量级、端到端可训练,并且可以轻松地插入到现有的 CNN 中,形成 IAUnet。实验表明,IAUnet 在图像和视频 reID 任务上均优于现有技术,并且在一般对象分类任务上也取得了引人注目的结果。源代码可在 https://github.com/blue-blue272/ImgReID-IAnet 上获得。

相似文献

1
IAUnet: Global Context-Aware Feature Learning for Person Reidentification.IAUnet:用于人体重识别的全局上下文感知特征学习。
IEEE Trans Neural Netw Learn Syst. 2021 Oct;32(10):4460-4474. doi: 10.1109/TNNLS.2020.3017939. Epub 2021 Oct 5.
2
Feature Completion for Occluded Person Re-Identification.遮挡行人再识别的特征补全。
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):4894-4912. doi: 10.1109/TPAMI.2021.3079910. Epub 2022 Aug 4.
3
Multi-scale Temporal Cues Learning for Video Person Re-Identification.用于视频人物重识别的多尺度时间线索学习
IEEE Trans Image Process. 2020 Feb 14. doi: 10.1109/TIP.2020.2972108.
4
TCGL: Temporal Contrastive Graph for Self-Supervised Video Representation Learning.TCGL:用于自监督视频表征学习的时间对比图
IEEE Trans Image Process. 2022;31:1978-1993. doi: 10.1109/TIP.2022.3147032. Epub 2022 Feb 18.
5
Spatially-Aware Context Neural Networks.空间感知上下文神经网络
IEEE Trans Image Process. 2021;30:6906-6916. doi: 10.1109/TIP.2021.3097917. Epub 2021 Aug 6.
6
Edge Preserving and Multi-Scale Contextual Neural Network for Salient Object Detection.边缘保持和多尺度上下文神经网络的显著目标检测。
IEEE Trans Image Process. 2018;27(1):121-134. doi: 10.1109/TIP.2017.2756825.
7
Global Aggregation Then Local Distribution for Scene Parsing.用于场景解析的全局聚合然后局部分布
IEEE Trans Image Process. 2021;30:6829-6842. doi: 10.1109/TIP.2021.3099366.
8
Towards Efficient Scene Understanding via Squeeze Reasoning.通过挤压推理实现高效场景理解。
IEEE Trans Image Process. 2021;30:7050-7063. doi: 10.1109/TIP.2021.3099369. Epub 2021 Aug 10.
9
Heterogeneous feature-aware Transformer-CNN coupling network for person re-identification.用于行人重识别的异构特征感知Transformer-CNN耦合网络
PeerJ Comput Sci. 2022 Sep 27;8:e1098. doi: 10.7717/peerj-cs.1098. eCollection 2022.
10
CTNet: Context-Based Tandem Network for Semantic Segmentation.CTNet:用于语义分割的基于上下文的串联网络
IEEE Trans Pattern Anal Mach Intell. 2022 Dec;44(12):9904-9917. doi: 10.1109/TPAMI.2021.3132068. Epub 2022 Nov 7.

引用本文的文献

1
Convolutional MLP orthogonal fusion of multiscale features for visual place recognition.用于视觉场所识别的多尺度特征卷积MLP正交融合
Sci Rep. 2024 May 23;14(1):11756. doi: 10.1038/s41598-024-62749-x.
2
AA-RGTCN: reciprocal global temporal convolution network with adaptive alignment for video-based person re-identification.AA-RGTCN:用于基于视频的行人重识别的具有自适应对齐的互逆全局时间卷积网络
Front Neurosci. 2024 Mar 25;18:1329884. doi: 10.3389/fnins.2024.1329884. eCollection 2024.
3
Siamese hierarchical feature fusion transformer for efficient tracking.
用于高效跟踪的暹罗层次特征融合变压器
Front Neurorobot. 2022 Dec 1;16:1082346. doi: 10.3389/fnbot.2022.1082346. eCollection 2022.