• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一款提供更多解决方案的视觉追踪器。

A Visual Tracker Offering More Solutions.

机构信息

College of Information and Computer Engineering, Northeast Forestry University, Harbin 150040, China.

Big Data Institute, East University of Heilongjiang, Harbin 150066, China.

出版信息

Sensors (Basel). 2020 Sep 19;20(18):5374. doi: 10.3390/s20185374.

DOI:10.3390/s20185374
PMID:32961752
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7570860/
Abstract

Most trackers focus solely on robustness and accuracy. Visual tracking, however, is a long-term problem with a high time limitation. A tracker that is robust, accurate, with long-term sustainability and real-time processing, is of high research value and practical significance. In this paper, we comprehensively consider these requirements in order to propose a new, state-of-the-art tracker with an excellent performance. EfficientNet-B0 is adopted for the first time via neural architecture search technology as the backbone network for the tracking task. This improves the network feature extraction ability and significantly reduces the number of parameters required for the tracker backbone network. In addition, maximal Distance Intersection-over-Union is set as the target estimation method, enhancing network stability and increasing the offline training convergence rate. Channel and spatial dual attention mechanisms are employed in the target classification module to improve the discrimination of the trackers. Furthermore, the conjugate gradient optimization strategy increases the speed of the online learning target classification module. A two-stage search method combined with a screening module is proposed to enable the tracker to cope with sudden target movement and reappearance following a brief disappearance. Our proposed method has an obvious speed advantage compared with pure global searching and achieves an optimal performance on OTB2015, VOT2016, VOT2018-LT, UAV-123 and LaSOT while running at over 50 FPS.

摘要

大多数跟踪器仅专注于鲁棒性和准确性。然而,视觉跟踪是一个长期存在的问题,时间限制很高。一个具有鲁棒性、准确性、长期可持续性和实时处理能力的跟踪器具有很高的研究价值和现实意义。在本文中,我们综合考虑了这些要求,提出了一种具有优异性能的新型、最先进的跟踪器。首次通过神经架构搜索技术采用 EfficientNet-B0 作为跟踪任务的骨干网络,这提高了网络的特征提取能力,并显著减少了跟踪器骨干网络所需的参数数量。此外,最大交并比被设置为目标估计方法,增强了网络稳定性,提高了离线训练的收敛速度。在目标分类模块中采用通道和空间双重注意力机制,提高了跟踪器的辨别能力。此外,共轭梯度优化策略提高了在线学习目标分类模块的速度。提出了一种两阶段搜索方法和筛选模块相结合的方法,使跟踪器能够应对目标突然运动和短暂消失后的再次出现。与纯全局搜索相比,我们提出的方法具有明显的速度优势,在 OTB2015、VOT2016、VOT2018-LT、UAV-123 和 LaSOT 上的性能达到最优,同时运行速度超过 50 FPS。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/3adc285e3cb9/sensors-20-05374-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/e7e4567d252f/sensors-20-05374-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/23ac0b4ceada/sensors-20-05374-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/429e0a47ac62/sensors-20-05374-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/5ac7876e6d89/sensors-20-05374-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/bef8e6e08a7d/sensors-20-05374-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/3943a96ed25a/sensors-20-05374-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/54fcb063a16c/sensors-20-05374-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/a998891f1a8f/sensors-20-05374-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/f1f949245771/sensors-20-05374-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/89f61b83d11a/sensors-20-05374-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/27b6575cefa7/sensors-20-05374-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/f729a69cef8d/sensors-20-05374-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/c958c8db95ca/sensors-20-05374-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/17dc9ea14903/sensors-20-05374-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/3adc285e3cb9/sensors-20-05374-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/e7e4567d252f/sensors-20-05374-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/23ac0b4ceada/sensors-20-05374-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/429e0a47ac62/sensors-20-05374-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/5ac7876e6d89/sensors-20-05374-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/bef8e6e08a7d/sensors-20-05374-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/3943a96ed25a/sensors-20-05374-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/54fcb063a16c/sensors-20-05374-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/a998891f1a8f/sensors-20-05374-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/f1f949245771/sensors-20-05374-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/89f61b83d11a/sensors-20-05374-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/27b6575cefa7/sensors-20-05374-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/f729a69cef8d/sensors-20-05374-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/c958c8db95ca/sensors-20-05374-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/17dc9ea14903/sensors-20-05374-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c9c4/7570860/3adc285e3cb9/sensors-20-05374-g014.jpg

相似文献

1
A Visual Tracker Offering More Solutions.一款提供更多解决方案的视觉追踪器。
Sensors (Basel). 2020 Sep 19;20(18):5374. doi: 10.3390/s20185374.
2
Robust Template Adjustment Siamese Network for Object Visual Tracking.用于目标视觉跟踪的鲁棒模板调整暹罗网络
Sensors (Basel). 2021 Feb 20;21(4):1466. doi: 10.3390/s21041466.
3
Siamese network with a depthwise over-parameterized convolutional layer for visual tracking.用于视觉跟踪的带有深度过度参数化卷积层的暹罗网络。
PLoS One. 2022 Aug 31;17(8):e0273690. doi: 10.1371/journal.pone.0273690. eCollection 2022.
4
Learning to Rank Proposals for Siamese Visual Tracking.学习对孪生视觉跟踪的提案进行排序。
IEEE Trans Image Process. 2021;30:8785-8796. doi: 10.1109/TIP.2021.3120305. Epub 2021 Oct 27.
5
EMAT: Efficient feature fusion network for visual tracking via optimized multi-head attention.EMAT:通过优化多头注意力进行视觉跟踪的高效特征融合网络。
Neural Netw. 2024 Apr;172:106110. doi: 10.1016/j.neunet.2024.106110. Epub 2024 Jan 6.
6
Improving Object Tracking by Added Noise and Channel Attention.添加噪声和通道注意力以改进目标跟踪。
Sensors (Basel). 2020 Jul 6;20(13):3780. doi: 10.3390/s20133780.
7
Appearance variation adaptation tracker using adversarial network.基于对抗网络的外观变化适应跟踪器。
Neural Netw. 2020 Sep;129:334-343. doi: 10.1016/j.neunet.2020.06.011. Epub 2020 Jun 17.
8
Learning Geometry Information of Target for Visual Object Tracking with Siamese Networks.利用连体网络学习视觉目标跟踪中目标的几何信息。
Sensors (Basel). 2021 Nov 23;21(23):7790. doi: 10.3390/s21237790.
9
TDIOT: Target-Driven Inference for Deep Video Object Tracking.TDIOT:深度视频目标跟踪的目标驱动推理
IEEE Trans Image Process. 2021;30:7938-7951. doi: 10.1109/TIP.2021.3112010. Epub 2021 Sep 22.
10
Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking.基于软掩模的特征融合与通道和空间注意力的鲁棒视觉目标跟踪
Sensors (Basel). 2020 Jul 20;20(14):4021. doi: 10.3390/s20144021.

引用本文的文献

1
Deep Learning for Detecting Subretinal Fluid and Discerning Macular Status by Fundus Images in Central Serous Chorioretinopathy.深度学习用于通过中心性浆液性脉络膜视网膜病变的眼底图像检测视网膜下液并识别黄斑状态
Front Bioeng Biotechnol. 2021 Nov 5;9:651340. doi: 10.3389/fbioe.2021.651340. eCollection 2021.
2
Aerial Video Trackers Review.航空视频跟踪器综述
Entropy (Basel). 2020 Nov 30;22(12):1358. doi: 10.3390/e22121358.

本文引用的文献

1
GOT-10k: A Large High-Diversity Benchmark for Generic Object Tracking in the Wild.GOT-10k:用于野外通用目标跟踪的大型高多样性基准数据集。
IEEE Trans Pattern Anal Mach Intell. 2021 May;43(5):1562-1577. doi: 10.1109/TPAMI.2019.2957464. Epub 2021 Apr 1.
2
Learning Support Correlation Filters for Visual Tracking.学习支持相关滤波器的视觉跟踪。
IEEE Trans Pattern Anal Mach Intell. 2019 May;41(5):1158-1172. doi: 10.1109/TPAMI.2018.2829180. Epub 2018 Apr 20.
3
Hedging Deep Features for Visual Tracking.基于深度特征的视觉跟踪的套期保值。
IEEE Trans Pattern Anal Mach Intell. 2019 May;41(5):1116-1130. doi: 10.1109/TPAMI.2018.2828817. Epub 2018 Apr 20.
4
High-Speed Tracking with Kernelized Correlation Filters.基于核相关滤波器的高速跟踪。
IEEE Trans Pattern Anal Mach Intell. 2015 Mar;37(3):583-96. doi: 10.1109/TPAMI.2014.2345390.
5
Object Tracking Benchmark.目标跟踪基准测试。
IEEE Trans Pattern Anal Mach Intell. 2015 Sep;37(9):1834-48. doi: 10.1109/TPAMI.2014.2388226.