• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于分层卷积特征的鲁棒视觉跟踪。

Robust Visual Tracking via Hierarchical Convolutional Features.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2019 Nov;41(11):2709-2723. doi: 10.1109/TPAMI.2018.2865311. Epub 2018 Aug 13.

DOI:10.1109/TPAMI.2018.2865311
PMID:30106709
Abstract

Visual tracking is challenging as target objects often undergo significant appearance changes caused by deformation, abrupt motion, background clutter and occlusion. In this paper, we propose to exploit the rich hierarchical features of deep convolutional neural networks to improve the accuracy and robustness of visual tracking. Deep neural networks trained on object recognition datasets consist of multiple convolutional layers. These layers encode target appearance with different levels of abstraction. For example, the outputs of the last convolutional layers encode the semantic information of targets and such representations are invariant to significant appearance variations. However, their spatial resolutions are too coarse to precisely localize the target. In contrast, features from earlier convolutional layers provide more precise localization but are less invariant to appearance changes. We interpret the hierarchical features of convolutional layers as a nonlinear counterpart of an image pyramid representation and explicitly exploit these multiple levels of abstraction to represent target objects. Specifically, we learn adaptive correlation filters on the outputs from each convolutional layer to encode the target appearance. We infer the maximum response of each layer to locate targets in a coarse-to-fine manner. To further handle the issues with scale estimation and re-detecting target objects from tracking failures caused by heavy occlusion or out-of-the-view movement, we conservatively learn another correlation filter, that maintains a long-term memory of target appearance, as a discriminative classifier. We apply the classifier to two types of object proposals: (1) proposals with a small step size and tightly around the estimated location for scale estimation; and (2) proposals with large step size and across the whole image for target re-detection. Extensive experimental results on large-scale benchmark datasets show that the proposed algorithm performs favorably against the state-of-the-art tracking methods.

摘要

视觉跟踪具有挑战性,因为目标对象通常会经历由于变形、突然运动、背景杂波和遮挡而引起的显著外观变化。在本文中,我们提出利用深度卷积神经网络的丰富层次特征来提高视觉跟踪的准确性和鲁棒性。在对象识别数据集上训练的深度神经网络由多个卷积层组成。这些层使用不同抽象级别的目标外观进行编码。例如,最后卷积层的输出编码目标的语义信息,并且这些表示对于显著的外观变化是不变的。然而,它们的空间分辨率太粗糙,无法精确地定位目标。相比之下,来自早期卷积层的特征提供了更精确的定位,但对外观变化的不变性较差。我们将卷积层的层次特征解释为图像金字塔表示的非线性对应物,并明确利用这些多个抽象级别来表示目标对象。具体来说,我们在每个卷积层的输出上学习自适应相关滤波器,以编码目标外观。我们以粗到精的方式推断每层的最大响应以定位目标。为了进一步处理由于严重遮挡或超出视野运动而导致的尺度估计和重新检测目标对象的问题,我们保守地学习另一个相关滤波器,该滤波器作为判别分类器,保持目标外观的长期记忆。我们将分类器应用于两种类型的对象建议:(1)具有小步长且紧密围绕估计位置的建议,用于尺度估计;(2)具有大步长且跨越整个图像的建议,用于目标重新检测。在大规模基准数据集上的广泛实验结果表明,所提出的算法在跟踪方法方面表现出色。

相似文献

1
Robust Visual Tracking via Hierarchical Convolutional Features.基于分层卷积特征的鲁棒视觉跟踪。
IEEE Trans Pattern Anal Mach Intell. 2019 Nov;41(11):2709-2723. doi: 10.1109/TPAMI.2018.2865311. Epub 2018 Aug 13.
2
Online Scale Adaptive Visual Tracking Based on Multilayer Convolutional Features.基于多层卷积特征的在线尺度自适应视觉跟踪。
IEEE Trans Cybern. 2019 Jan;49(1):146-158. doi: 10.1109/TCYB.2017.2768570. Epub 2017 Nov 14.
3
Hierarchical Recurrent Neural Hashing for Image Retrieval With Hierarchical Convolutional Features.基于层次卷积特征的层次递归神经网络哈希图像检索
IEEE Trans Image Process. 2018;27(1):106-120. doi: 10.1109/TIP.2017.2755766.
4
Hedging Deep Features for Visual Tracking.基于深度特征的视觉跟踪的套期保值。
IEEE Trans Pattern Anal Mach Intell. 2019 May;41(5):1116-1130. doi: 10.1109/TPAMI.2018.2828817. Epub 2018 Apr 20.
5
Robust Visual Tracking via Convolutional Networks Without Training.基于卷积网络的无需训练的鲁棒视觉跟踪。
IEEE Trans Image Process. 2016 Apr;25(4):1779-92. doi: 10.1109/TIP.2016.2531283. Epub 2016 Feb 18.
6
Video tracking using learned hierarchical features.基于学习的分层特征的视频跟踪。
IEEE Trans Image Process. 2015 Apr;24(4):1424-35. doi: 10.1109/TIP.2015.2403231. Epub 2015 Feb 12.
7
Dual-regression model for visual tracking.基于回归的视觉跟踪模型。
Neural Netw. 2020 Dec;132:364-374. doi: 10.1016/j.neunet.2020.09.011. Epub 2020 Sep 24.
8
Deep Attention Models for Human Tracking Using RGBD.基于 RGBD 的深度注意模型的人体跟踪
Sensors (Basel). 2019 Feb 13;19(4):750. doi: 10.3390/s19040750.
9
Deformable Object Tracking With Gated Fusion.基于门控融合的可变形物体跟踪
IEEE Trans Image Process. 2019 Aug;28(8):3766-3777. doi: 10.1109/TIP.2019.2902784. Epub 2019 Mar 7.
10
Visual Tracking via Coarse and Fine Structural Local Sparse Appearance Models.基于粗粒度和细粒度结构局部稀疏表观模型的视觉跟踪。
IEEE Trans Image Process. 2016 Oct;25(10):4555-64. doi: 10.1109/TIP.2016.2592701. Epub 2016 Jul 18.

引用本文的文献

1
DepthCrackNet: A Deep Learning Model for Automatic Pavement Crack Detection.深度裂缝网络:一种用于自动路面裂缝检测的深度学习模型。
J Imaging. 2024 Apr 26;10(5):100. doi: 10.3390/jimaging10050100.
2
Anchor free based Siamese network tracker with transformer for RGB-T tracking.基于无锚点暹罗网络的带有Transformer的RGB-T跟踪器
Sci Rep. 2023 Aug 16;13(1):13294. doi: 10.1038/s41598-023-39978-7.
3
An adaptive spatiotemporal correlation filtering visual tracking method.一种自适应时空相关滤波视觉跟踪方法。
PLoS One. 2023 Jan 6;18(1):e0279240. doi: 10.1371/journal.pone.0279240. eCollection 2023.
4
Shape-Texture Debiased Training for Robust Template Matching.形状-纹理去偏训练用于鲁棒模板匹配。
Sensors (Basel). 2022 Sep 2;22(17):6658. doi: 10.3390/s22176658.
5
Multi-Feature Single Target Robust Tracking Fused with Particle Filter.融合粒子滤波器的多特征单目标鲁棒跟踪
Sensors (Basel). 2022 Feb 27;22(5):1879. doi: 10.3390/s22051879.
6
Visual Tracking via Deep Feature Fusion and Correlation Filters.基于深度特征融合与相关滤波的视觉跟踪
Sensors (Basel). 2020 Jun 14;20(12):3370. doi: 10.3390/s20123370.
7
Online Siamese Network for Visual Object Tracking.用于视觉目标跟踪的在线暹罗网络
Sensors (Basel). 2019 Apr 18;19(8):1858. doi: 10.3390/s19081858.
8
Distractor-Aware Deep Regression for Visual Tracking.基于分心物感知的深度回归视觉跟踪方法
Sensors (Basel). 2019 Jan 18;19(2):387. doi: 10.3390/s19020387.
9
A 3D Relative-Motion Context Constraint-Based MAP Solution for Multiple-Object Tracking Problems.基于三维相对运动上下文约束的 MAP 解算的多目标跟踪问题。
Sensors (Basel). 2018 Jul 20;18(7):2363. doi: 10.3390/s18072363.
10
Deep Auto-context Convolutional Neural Networks for Standard-Dose PET Image Estimation from Low-Dose PET/MRI.用于从低剂量PET/MRI估计标准剂量PET图像的深度自动上下文卷积神经网络
Neurocomputing (Amst). 2017 Dec 6;267:406-416. doi: 10.1016/j.neucom.2017.06.048. Epub 2017 Jun 29.