• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于软掩模的特征融合与通道和空间注意力的鲁棒视觉目标跟踪

Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking.

机构信息

School of Computer Science and Engineering, Kyungpook National University, Daegu 41566, Korea.

Department of Computer Science, Information Technology University, Lahore 54000, Pakistan.

出版信息

Sensors (Basel). 2020 Jul 20;20(14):4021. doi: 10.3390/s20144021.

DOI:10.3390/s20144021
PMID:32698339
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7412361/
Abstract

We propose to improve the visual object tracking by introducing a soft mask based low-level feature fusion technique. The proposed technique is further strengthened by integrating channel and spatial attention mechanisms. The proposed approach is integrated within a Siamese framework to demonstrate its effectiveness for visual object tracking. The proposed soft mask is used to give more importance to the target regions as compared to the other regions to enable effective target feature representation and to increase discriminative power. The low-level feature fusion improves the tracker robustness against distractors. The channel attention is used to identify more discriminative channels for better target representation. The spatial attention complements the soft mask based approach to better localize the target objects in challenging tracking scenarios. We evaluated our proposed approach over five publicly available benchmark datasets and performed extensive comparisons with 39 state-of-the-art tracking algorithms. The proposed tracker demonstrates excellent performance compared to the existing state-of-the-art trackers.

摘要

我们提出了一种基于软掩模的底层特征融合技术来改进视觉目标跟踪。该技术通过集成通道和空间注意力机制得到进一步加强。所提出的方法集成在一个孪生框架内,以证明其在视觉目标跟踪方面的有效性。所提出的软掩模用于与其他区域相比,给予目标区域更多的重要性,从而实现有效的目标特征表示并提高判别力。底层特征融合提高了跟踪器对干扰物的鲁棒性。通道注意力用于为更好的目标表示识别更具判别力的通道。空间注意力补充了基于软掩模的方法,以便在具有挑战性的跟踪场景中更好地定位目标对象。我们在五个公开可用的基准数据集上评估了我们提出的方法,并与 39 种最先进的跟踪算法进行了广泛的比较。与现有的最先进的跟踪器相比,所提出的跟踪器表现出了优异的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/868fd5b0b6e3/sensors-20-04021-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/94cfb6e537ef/sensors-20-04021-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/3e10f7e40e49/sensors-20-04021-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/852f20775529/sensors-20-04021-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/2162586a3be9/sensors-20-04021-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/acc3b825b656/sensors-20-04021-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/1bac51434a05/sensors-20-04021-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/282fbab0cefe/sensors-20-04021-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/6b379877e338/sensors-20-04021-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/5ffe75ca6600/sensors-20-04021-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/868fd5b0b6e3/sensors-20-04021-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/94cfb6e537ef/sensors-20-04021-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/3e10f7e40e49/sensors-20-04021-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/852f20775529/sensors-20-04021-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/2162586a3be9/sensors-20-04021-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/acc3b825b656/sensors-20-04021-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/1bac51434a05/sensors-20-04021-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/282fbab0cefe/sensors-20-04021-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/6b379877e338/sensors-20-04021-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/5ffe75ca6600/sensors-20-04021-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1e4e/7412361/868fd5b0b6e3/sensors-20-04021-g010.jpg

相似文献

1
Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking.基于软掩模的特征融合与通道和空间注意力的鲁棒视觉目标跟踪
Sensors (Basel). 2020 Jul 20;20(14):4021. doi: 10.3390/s20144021.
2
Improving Object Tracking by Added Noise and Channel Attention.添加噪声和通道注意力以改进目标跟踪。
Sensors (Basel). 2020 Jul 6;20(13):3780. doi: 10.3390/s20133780.
3
HROM: Learning High-Resolution Representation and Object-Aware Masks for Visual Object Tracking.HROM:用于视觉目标跟踪的学习高分辨率表示和对象感知掩模。
Sensors (Basel). 2020 Aug 26;20(17):4807. doi: 10.3390/s20174807.
4
Learning Deep Lucas-Kanade Siamese Network for Visual Tracking.用于视觉跟踪的深度卢卡斯-卡纳德连体网络学习
IEEE Trans Image Process. 2021;30:4814-4827. doi: 10.1109/TIP.2021.3076272. Epub 2021 May 7.
5
Siamese Implicit Region Proposal Network With Compound Attention for Visual Tracking.基于复合注意力的暹罗隐式区域提案网络的视觉跟踪。
IEEE Trans Image Process. 2022;31:1882-1894. doi: 10.1109/TIP.2022.3148876. Epub 2022 Feb 16.
6
Hedging Deep Features for Visual Tracking.基于深度特征的视觉跟踪的套期保值。
IEEE Trans Pattern Anal Mach Intell. 2019 May;41(5):1116-1130. doi: 10.1109/TPAMI.2018.2828817. Epub 2018 Apr 20.
7
SiamHAS: Siamese Tracker with Hierarchical Attention Strategy for Aerial Tracking.暹罗HAS:用于空中跟踪的具有分层注意力策略的暹罗跟踪器。
Micromachines (Basel). 2023 Apr 21;14(4):893. doi: 10.3390/mi14040893.
8
SGAT: Shuffle and graph attention based Siamese networks for visual tracking.基于 Shuffle 和图注意力的孪生网络的视觉跟踪。
PLoS One. 2022 Nov 23;17(11):e0277064. doi: 10.1371/journal.pone.0277064. eCollection 2022.
9
Siamese anchor-free object tracking with multiscale spatial attentions.基于多尺度空间注意力的暹罗无锚点目标跟踪
Sci Rep. 2021 Nov 25;11(1):22908. doi: 10.1038/s41598-021-02095-4.
10
HKSiamFC: Visual-Tracking Framework Using Prior Information Provided by Staple and Kalman Filter.HKSiamFC:基于 Staple 和卡尔曼滤波器提供的先验信息的视觉跟踪框架。
Sensors (Basel). 2020 Apr 10;20(7):2137. doi: 10.3390/s20072137.

引用本文的文献

1
Proposal-Based Visual Tracking Using Spatial Cascaded Transformed Region Proposal Network.基于提案的视觉跟踪:使用空间级联变换区域提案网络。
Sensors (Basel). 2020 Aug 26;20(17):4810. doi: 10.3390/s20174810.

本文引用的文献

1
Improving Object Tracking by Added Noise and Channel Attention.添加噪声和通道注意力以改进目标跟踪。
Sensors (Basel). 2020 Jul 6;20(13):3780. doi: 10.3390/s20133780.
2
GOT-10k: A Large High-Diversity Benchmark for Generic Object Tracking in the Wild.GOT-10k:用于野外通用目标跟踪的大型高多样性基准数据集。
IEEE Trans Pattern Anal Mach Intell. 2021 May;43(5):1562-1577. doi: 10.1109/TPAMI.2019.2957464. Epub 2021 Apr 1.
3
Visual Object Tracking by Hierarchical Attention Siamese Network.基于层次注意力暹罗网络的视觉目标跟踪
IEEE Trans Cybern. 2020 Jul;50(7):3068-3080. doi: 10.1109/TCYB.2019.2936503. Epub 2019 Sep 12.
4
Visual Tracking via Dynamic Memory Networks.通过动态记忆网络进行视觉跟踪。
IEEE Trans Pattern Anal Mach Intell. 2021 Jan;43(1):360-374. doi: 10.1109/TPAMI.2019.2929034. Epub 2020 Dec 4.
5
Learning to Update for Object Tracking with Recurrent Meta-learner.使用循环元学习器进行目标跟踪的更新学习
IEEE Trans Image Process. 2019 Feb 21. doi: 10.1109/TIP.2019.2900577.
6
Object Tracking Algorithm Based on Dual Color Feature Fusion with Dimension Reduction.基于降维的双通道特征融合的目标跟踪算法。
Sensors (Basel). 2018 Dec 25;19(1):73. doi: 10.3390/s19010073.
7
Visual Tracking via Dynamic Graph Learning.基于动态图学习的视觉跟踪
IEEE Trans Pattern Anal Mach Intell. 2019 Nov;41(11):2770-2782. doi: 10.1109/TPAMI.2018.2864965. Epub 2018 Aug 13.
8
HyperFace: A Deep Multi-Task Learning Framework for Face Detection, Landmark Localization, Pose Estimation, and Gender Recognition.超人脸:一个用于人脸检测、地标定位、姿势估计和性别识别的深度多任务学习框架。
IEEE Trans Pattern Anal Mach Intell. 2019 Jan;41(1):121-135. doi: 10.1109/TPAMI.2017.2781233. Epub 2017 Dec 8.
9
Deformable Parts Correlation Filters for Robust Visual Tracking.可变形部件相关滤波器用于鲁棒视觉跟踪。
IEEE Trans Cybern. 2018 Jun;48(6):1849-1861. doi: 10.1109/TCYB.2017.2716101. Epub 2017 Jun 27.
10
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.