• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

暹罗CAN:基于暹罗中心感知网络的实时视觉跟踪

SiamCAN: Real-Time Visual Tracking Based on Siamese Center-Aware Network.

作者信息

Zhou Wenzhang, Wen Longyin, Zhang Libo, Du Dawei, Luo Tiejian, Wu Yanjun

出版信息

IEEE Trans Image Process. 2021;30:3597-3609. doi: 10.1109/TIP.2021.3060905. Epub 2021 Mar 17.

DOI:10.1109/TIP.2021.3060905
PMID:33656991
Abstract

In this article, we present a novel Siamese center-aware network (SiamCAN) for visual tracking, which consists of the Siamese feature extraction subnetwork, followed by the classification, regression, and localization branches in parallel. The classification branch is used to distinguish the target from background, and the regression branch is introduced to regress the bounding box of the target. To reduce the impact of manually designed anchor boxes to adapt to different target motion patterns, we design the localization branch to localize the target center directly to assist the regression branch generating accurate results. Meanwhile, we introduce the global context module into the localization branch to capture long-range dependencies for more robustness to large displacements of the target. A multi-scale learnable attention module is used to guide these three branches to exploit discriminative features for better performance. Extensive experiments on 9 challenging benchmarks, namely VOT2016, VOT2018, VOT2019, OTB100, LTB35, LaSOT, TC128, UAV123 and VisDrone-SOT2019 demonstrate that SiamCAN achieves leading accuracy with high efficiency. Our source code is available at https://isrc.iscas.ac.cn/gitlab/research/siamcan.

摘要

在本文中,我们提出了一种用于视觉跟踪的新型暹罗中心感知网络(SiamCAN),它由暹罗特征提取子网络组成,随后并行地连接分类、回归和定位分支。分类分支用于区分目标与背景,回归分支用于回归目标的边界框。为了减少手动设计的锚框的影响以适应不同的目标运动模式,我们设计定位分支直接定位目标中心,以辅助回归分支生成准确结果。同时,我们将全局上下文模块引入定位分支,以捕捉远距离依赖关系,从而对目标的大位移具有更强的鲁棒性。使用多尺度可学习注意力模块来引导这三个分支利用判别性特征以获得更好的性能。在9个具有挑战性的基准测试上进行的大量实验,即VOT2016、VOT2018、VOT2019、OTB100、LTB35、LaSOT、TC128、UAV123和VisDrone-SOT2019,表明SiamCAN以高效率实现了领先的精度。我们的源代码可在https://isrc.iscas.ac.cn/gitlab/research/siamcan获取。

相似文献

1
SiamCAN: Real-Time Visual Tracking Based on Siamese Center-Aware Network.暹罗CAN:基于暹罗中心感知网络的实时视觉跟踪
IEEE Trans Image Process. 2021;30:3597-3609. doi: 10.1109/TIP.2021.3060905. Epub 2021 Mar 17.
2
SiamBAN: Target-Aware Tracking With Siamese Box Adaptive Network.暹罗边界感知网络:基于暹罗框自适应网络的目标感知跟踪
IEEE Trans Pattern Anal Mach Intell. 2023 Apr;45(4):5158-5173. doi: 10.1109/TPAMI.2022.3195759. Epub 2023 Mar 7.
3
Robust Template Adjustment Siamese Network for Object Visual Tracking.用于目标视觉跟踪的鲁棒模板调整暹罗网络
Sensors (Basel). 2021 Feb 20;21(4):1466. doi: 10.3390/s21041466.
4
HROM: Learning High-Resolution Representation and Object-Aware Masks for Visual Object Tracking.HROM:用于视觉目标跟踪的学习高分辨率表示和对象感知掩模。
Sensors (Basel). 2020 Aug 26;20(17):4807. doi: 10.3390/s20174807.
5
Toward Robust Visual Object Tracking With Independent Target-Agnostic Detection and Effective Siamese Cross-Task Interaction.通过独立的目标无关检测和有效的暹罗跨任务交互实现鲁棒视觉目标跟踪
IEEE Trans Image Process. 2023;32:1541-1554. doi: 10.1109/TIP.2023.3246800. Epub 2023 Mar 6.
6
Siamese anchor-free object tracking with multiscale spatial attentions.基于多尺度空间注意力的暹罗无锚点目标跟踪
Sci Rep. 2021 Nov 25;11(1):22908. doi: 10.1038/s41598-021-02095-4.
7
Learning Geometry Information of Target for Visual Object Tracking with Siamese Networks.利用连体网络学习视觉目标跟踪中目标的几何信息。
Sensors (Basel). 2021 Nov 23;21(23):7790. doi: 10.3390/s21237790.
8
Learning to Rank Proposals for Siamese Visual Tracking.学习对孪生视觉跟踪的提案进行排序。
IEEE Trans Image Process. 2021;30:8785-8796. doi: 10.1109/TIP.2021.3120305. Epub 2021 Oct 27.
9
Siamese network with a depthwise over-parameterized convolutional layer for visual tracking.用于视觉跟踪的带有深度过度参数化卷积层的暹罗网络。
PLoS One. 2022 Aug 31;17(8):e0273690. doi: 10.1371/journal.pone.0273690. eCollection 2022.
10
Adaptive ensemble perception tracking.自适应集成感知跟踪。
Neural Netw. 2021 Oct;142:316-328. doi: 10.1016/j.neunet.2021.05.003. Epub 2021 May 15.

引用本文的文献

1
AMST: aggregated multi-level spatial and temporal context-based transformer for robust aerial tracking.基于聚合多层次时空上下文的Transformer 模型用于稳健的空中目标跟踪
Sci Rep. 2023 Jun 4;13(1):9062. doi: 10.1038/s41598-023-36131-2.