• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

超越贪婪搜索:基于多智能体强化学习的束搜索跟踪

Beyond Greedy Search: Tracking by Multi-Agent Reinforcement Learning-Based Beam Search.

作者信息

Wang Xiao, Chen Zhe, Jiang Bo, Tang Jin, Luo Bin, Tao Dacheng

出版信息

IEEE Trans Image Process. 2022;31:6239-6254. doi: 10.1109/TIP.2022.3208437. Epub 2022 Sep 30.

DOI:10.1109/TIP.2022.3208437
PMID:36166563
Abstract

To track the target in a video, current visual trackers usually adopt greedy search for target object localization in each frame, that is, the candidate region with the maximum response score will be selected as the tracking result of each frame. However, we found that this may be not an optimal choice, especially when encountering challenging tracking scenarios such as heavy occlusion and fast motion. In particular, if a tracker drifts, errors will be accumulated and would further make response scores estimated by the tracker unreliable in future frames. To address this issue, we propose to maintain multiple tracking trajectories and apply beam search strategy for visual tracking, so that the trajectory with fewer accumulated errors can be identified. Accordingly, this paper introduces a novel multi-agent reinforcement learning based beam search tracking strategy, termed BeamTracking. It is mainly inspired by the image captioning task, which takes an image as input and generates diverse descriptions using beam search algorithm. Accordingly, we formulate the tracking as a sample selection problem fulfilled by multiple parallel decision-making processes, each of which aims at picking out one sample as their tracking result in each frame. Each maintained trajectory is associated with an agent to perform the decision-making and determine what actions should be taken to update related information. More specifically, using the classification-based tracker as the baseline, we first adopt bi-GRU to encode the target feature, proposal feature, and its response score into a unified state representation. The state feature and greedy search result are then fed into the first agent for independent action selection. Afterwards, the output action and state features are fed into the subsequent agent for diverse results prediction. When all the frames are processed, we select the trajectory with the maximum accumulated score as the tracking result. Extensive experiments on seven popular tracking benchmark datasets validated the effectiveness of the proposed algorithm.

摘要

为了在视频中跟踪目标,当前的视觉跟踪器通常在每一帧中采用贪婪搜索来定位目标对象,即选择响应得分最高的候选区域作为每一帧的跟踪结果。然而,我们发现这可能不是一个最优选择,尤其是在遇到诸如严重遮挡和快速运动等具有挑战性的跟踪场景时。特别是,如果跟踪器发生漂移,误差将会累积,并会进一步使跟踪器在未来帧中估计的响应得分变得不可靠。为了解决这个问题,我们建议维护多个跟踪轨迹,并将波束搜索策略应用于视觉跟踪,以便能够识别累积误差较少的轨迹。因此,本文介绍了一种基于多智能体强化学习的新型波束搜索跟踪策略,称为BeamTracking。它主要受到图像字幕任务的启发,该任务以图像为输入,并使用波束搜索算法生成多样化的描述。相应地,我们将跟踪问题表述为一个由多个并行决策过程完成的样本选择问题,每个决策过程旨在在每一帧中挑选出一个样本作为其跟踪结果。每个维护的轨迹都与一个智能体相关联,以执行决策并确定应采取哪些行动来更新相关信息。更具体地说,以基于分类的跟踪器为基线,我们首先采用双向门控循环单元(bi-GRU)将目标特征、提议特征及其响应得分编码为统一的状态表示。然后,将状态特征和贪婪搜索结果输入到第一个智能体中进行独立的动作选择。之后,将输出动作和状态特征输入到后续智能体中进行多样化结果预测。当所有帧都处理完毕后,我们选择累积得分最高的轨迹作为跟踪结果。在七个流行的跟踪基准数据集上进行的大量实验验证了所提算法的有效性。

相似文献

1
Beyond Greedy Search: Tracking by Multi-Agent Reinforcement Learning-Based Beam Search.超越贪婪搜索:基于多智能体强化学习的束搜索跟踪
IEEE Trans Image Process. 2022;31:6239-6254. doi: 10.1109/TIP.2022.3208437. Epub 2022 Sep 30.
2
Tracking by Joint Local and Global Search: A Target-Aware Attention-Based Approach.联合局部和全局搜索的跟踪:一种目标感知的注意力基础方法。
IEEE Trans Neural Netw Learn Syst. 2022 Nov;33(11):6931-6945. doi: 10.1109/TNNLS.2021.3083933. Epub 2022 Oct 27.
3
Adaptive Object Tracking via Multi-Angle Analysis Collaboration.基于多角度分析协作的自适应目标跟踪。
Sensors (Basel). 2018 Oct 24;18(11):3606. doi: 10.3390/s18113606.
4
SAFS: Object Tracking Algorithm Based on Self-Adaptive Feature Selection.SAFS:基于自适应特征选择的目标跟踪算法
Sensors (Basel). 2021 Jun 11;21(12):4030. doi: 10.3390/s21124030.
5
Motion-Aware Correlation Filters for Online Visual Tracking.运动感知相关滤波器的在线视觉跟踪。
Sensors (Basel). 2018 Nov 14;18(11):3937. doi: 10.3390/s18113937.
6
Robust Visual Tracking via Sparsity-Induced Subspace Learning.基于稀疏诱导子空间学习的鲁棒视觉跟踪。
IEEE Trans Image Process. 2015 Dec;24(12):4686-700. doi: 10.1109/TIP.2015.2462076. Epub 2015 Jul 30.
7
Modeling of Multiple Spatial-Temporal Relations for Robust Visual Object Tracking.用于鲁棒视觉目标跟踪的多时空关系建模
IEEE Trans Image Process. 2024;33:5073-5085. doi: 10.1109/TIP.2024.3453028. Epub 2024 Sep 17.
8
Correlation-Based Tracker-Level Fusion for Robust Visual Tracking.基于相关的跟踪器级融合用于鲁棒视觉跟踪。
IEEE Trans Image Process. 2017 Oct;26(10):4832-4842. doi: 10.1109/TIP.2017.2699791. Epub 2017 Apr 28.
9
Jointly Feature Learning and Selection for Robust Tracking via a Gating Mechanism.通过门控机制进行鲁棒跟踪的联合特征学习与选择
PLoS One. 2016 Aug 30;11(8):e0161808. doi: 10.1371/journal.pone.0161808. eCollection 2016.
10
Hierarchical Tracking by Reinforcement Learning based Searching and Coarse-to-fine Verifying.基于强化学习搜索和粗到精验证的分层跟踪
IEEE Trans Image Process. 2018 Dec 5. doi: 10.1109/TIP.2018.2885238.

引用本文的文献

1
Introducing π-HelixNovo for practical large-scale de novo peptide sequencing.介绍用于实际大规模从头多肽测序的 π-螺旋 Novo。
Brief Bioinform. 2024 Jan 22;25(2). doi: 10.1093/bib/bbae021.
2
Optimized glycemic control of type 2 diabetes with reinforcement learning: a proof-of-concept trial.基于强化学习的 2 型糖尿病优化血糖控制:一项概念验证试验。
Nat Med. 2023 Oct;29(10):2633-2642. doi: 10.1038/s41591-023-02552-9. Epub 2023 Sep 14.