• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于收缩损失的深度目标跟踪

Deep Object Tracking With Shrinkage Loss.

作者信息

Lu Xiankai, Ma Chao, Shen Jianbing, Yang Xiaokang, Reid Ian, Yang Ming-Hsuan

出版信息

IEEE Trans Pattern Anal Mach Intell. 2022 May;44(5):2386-2401. doi: 10.1109/TPAMI.2020.3041332. Epub 2022 Apr 1.

DOI:10.1109/TPAMI.2020.3041332
PMID:33253114
Abstract

In this paper, we address the issue of data imbalance in learning deep models for visual object tracking. Although it is well known that data distribution plays a crucial role in learning and inference models, considerably less attention has been paid to data imbalance in visual tracking. For the deep regression trackers that directly learn a dense mapping from input images of target objects to soft response maps, we identify their performance is limited by the extremely imbalanced pixel-to-pixel differences when computing regression loss. This prevents existing end-to-end learnable deep regression trackers from performing as well as discriminative correlation filters (DCFs) trackers. For the deep classification trackers that draw positive and negative samples to learn discriminative classifiers, there exists heavy class imbalance due to a limited number of positive samples when compared to the number of negative samples. To balance training data, we propose a novel shrinkage loss to penalize the importance of easy training data mostly coming from the background, which facilitates both deep regression and classification trackers to better distinguish target objects from the background. We extensively validate the proposed shrinkage loss function on six benchmark datasets, including the OTB-2013, OTB-2015, UAV-123, VOT-2016, VOT-2018 and LaSOT. Equipped with our shrinkage loss, the proposed one-stage deep regression tracker achieves favorable results against state-of-the-art methods, especially in comparison with DCFs trackers. Meanwhile, our shrinkage loss generalizes well to deep classification trackers. When replacing the original binary cross entropy loss with our shrinkage loss, three representative baseline trackers achieve large performance gains, even setting new state-of-the-art results.

摘要

在本文中,我们探讨了用于视觉目标跟踪的深度模型学习中的数据不平衡问题。尽管众所周知数据分布在学习和推理模型中起着至关重要的作用,但视觉跟踪中的数据不平衡问题却很少受到关注。对于直接从目标对象的输入图像学习到软响应图的深度回归跟踪器,我们发现其性能受到计算回归损失时像素间极度不平衡差异的限制。这使得现有的端到端可学习深度回归跟踪器无法达到判别相关滤波器(DCF)跟踪器的性能。对于通过抽取正负样本学习判别分类器的深度分类跟踪器,与负样本数量相比,由于正样本数量有限,存在严重的类别不平衡。为了平衡训练数据,我们提出了一种新颖的收缩损失,以惩罚主要来自背景的简单训练数据的重要性,这有助于深度回归和分类跟踪器更好地将目标对象与背景区分开来。我们在六个基准数据集上广泛验证了所提出的收缩损失函数,包括OTB - 2013、OTB - 2015、UAV - 123、VOT - 2016、VOT - 2018和LaSOT。配备了我们的收缩损失,所提出的单阶段深度回归跟踪器相对于现有方法取得了良好的结果,特别是与DCF跟踪器相比。同时,我们的收缩损失在深度分类跟踪器中也具有良好的通用性。当用我们的收缩损失替换原始的二元交叉熵损失时,三个有代表性的基线跟踪器取得了显著的性能提升,甚至创造了新的最优结果。

相似文献

1
Deep Object Tracking With Shrinkage Loss.基于收缩损失的深度目标跟踪
IEEE Trans Pattern Anal Mach Intell. 2022 May;44(5):2386-2401. doi: 10.1109/TPAMI.2020.3041332. Epub 2022 Apr 1.
2
Distractor-Aware Deep Regression for Visual Tracking.基于分心物感知的深度回归视觉跟踪方法
Sensors (Basel). 2019 Jan 18;19(2):387. doi: 10.3390/s19020387.
3
Proposal-Based Visual Tracking Using Spatial Cascaded Transformed Region Proposal Network.基于提案的视觉跟踪:使用空间级联变换区域提案网络。
Sensors (Basel). 2020 Aug 26;20(17):4810. doi: 10.3390/s20174810.
4
Motion-Aware Correlation Filters for Online Visual Tracking.运动感知相关滤波器的在线视觉跟踪。
Sensors (Basel). 2018 Nov 14;18(11):3937. doi: 10.3390/s18113937.
5
Hierarchical Spatiotemporal Context-Aware Correlation Filters for Visual Tracking.分层时空上下文感知相关滤波器的视觉跟踪。
IEEE Trans Cybern. 2021 Dec;51(12):6066-6079. doi: 10.1109/TCYB.2020.2964757. Epub 2021 Dec 22.
6
Siamese Implicit Region Proposal Network With Compound Attention for Visual Tracking.基于复合注意力的暹罗隐式区域提案网络的视觉跟踪。
IEEE Trans Image Process. 2022;31:1882-1894. doi: 10.1109/TIP.2022.3148876. Epub 2022 Feb 16.
7
Good Features to Correlate for Visual Tracking.用于视觉跟踪的相关优良特征。
IEEE Trans Image Process. 2018 Feb 14. doi: 10.1109/TIP.2018.2806280.
8
Benchmarking Deep Trackers on Aerial Videos.在航空视频上进行深度跟踪器的基准测试。
Sensors (Basel). 2020 Jan 19;20(2):547. doi: 10.3390/s20020547.
9
A practical evaluation of correlation filter-based object trackers with new features.基于相关滤波的具有新特征的目标跟踪器的实用评估。
PLoS One. 2022 Aug 25;17(8):e0273022. doi: 10.1371/journal.pone.0273022. eCollection 2022.
10
Hedging Deep Features for Visual Tracking.基于深度特征的视觉跟踪的套期保值。
IEEE Trans Pattern Anal Mach Intell. 2019 May;41(5):1116-1130. doi: 10.1109/TPAMI.2018.2828817. Epub 2018 Apr 20.

引用本文的文献

1
A new model for classification of medical CT images using CNN: a COVID-19 case study.一种使用卷积神经网络进行医学CT图像分类的新模型:以COVID-19为例的研究。
Multimed Tools Appl. 2022 Dec 19:1-29. doi: 10.1007/s11042-022-14316-7.