• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

端到端主动目标跟踪及其通过强化学习的实际部署。

End-to-End Active Object Tracking and Its Real-World Deployment via Reinforcement Learning.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2020 Jun;42(6):1317-1332. doi: 10.1109/TPAMI.2019.2899570. Epub 2019 Feb 14.

DOI:10.1109/TPAMI.2019.2899570
PMID:30762532
Abstract

We study active object tracking, where a tracker takes visual observations (i.e., frame sequences) as input and produces the corresponding camera control signals as output (e.g., move forward, turn left, etc.). Conventional methods tackle tracking and camera control tasks separately, and the resulting system is difficult to tune jointly. These methods also require significant human efforts for image labeling and expensive trial-and-error system tuning in the real world. To address these issues, we propose, in this paper, an end-to-end solution via deep reinforcement learning. A ConvNet-LSTM function approximator is adopted for the direct frame-to-action prediction. We further propose an environment augmentation technique and a customized reward function, which are crucial for successful training. The tracker trained in simulators (ViZDoom and Unreal Engine) demonstrates good generalization behaviors in the case of unseen object moving paths, unseen object appearances, unseen backgrounds, and distracting objects. The system is robust and can restore tracking after occasional lost of the target being tracked. We also find that the tracking ability, obtained solely from simulators, can potentially transfer to real-world scenarios. We demonstrate successful examples of such transfer, via experiments over the VOT dataset and the deployment of a real-world robot using the proposed active tracker trained in simulation.

摘要

我们研究主动目标跟踪,其中跟踪器以视觉观察(即帧序列)作为输入,并生成相应的相机控制信号作为输出(例如,前进、左转等)。传统方法分别解决跟踪和相机控制任务,并且难以联合调整由此产生的系统。这些方法还需要大量的人力进行图像标记,并且在现实世界中需要昂贵的反复试验系统调整。为了解决这些问题,我们在本文中提出了一种通过深度强化学习的端到端解决方案。采用 ConvNet-LSTM 函数逼近器进行直接帧到动作预测。我们进一步提出了一种环境增强技术和定制的奖励函数,这对于成功训练至关重要。在看不见的目标运动路径、看不见的目标外观、看不见的背景和干扰物体的情况下,在模拟器(ViZDoom 和 Unreal Engine)中训练的跟踪器表现出良好的泛化行为。该系统具有鲁棒性,可以在偶尔丢失跟踪目标后恢复跟踪。我们还发现,仅从模拟器获得的跟踪能力可能会潜在地转移到真实场景中。我们通过在 VOT 数据集上进行的实验以及使用在模拟中训练的主动跟踪器在真实机器人上的部署,展示了这种转移的成功示例。

相似文献

1
End-to-End Active Object Tracking and Its Real-World Deployment via Reinforcement Learning.端到端主动目标跟踪及其通过强化学习的实际部署。
IEEE Trans Pattern Anal Mach Intell. 2020 Jun;42(6):1317-1332. doi: 10.1109/TPAMI.2019.2899570. Epub 2019 Feb 14.
2
Trajectory Predictor by Using Recurrent Neural Networks in Visual Tracking.基于循环神经网络的视觉跟踪轨迹预测器
IEEE Trans Cybern. 2017 Oct;47(10):3172-3183. doi: 10.1109/TCYB.2017.2705345.
3
Action-Driven Visual Object Tracking With Deep Reinforcement Learning.基于深度强化学习的驱动式视觉目标跟踪
IEEE Trans Neural Netw Learn Syst. 2018 Jun;29(6):2239-2252. doi: 10.1109/TNNLS.2018.2801826.
4
AD-VAT+: An Asymmetric Dueling Mechanism for Learning and Understanding Visual Active Tracking.AD-VAT+: 一种用于学习和理解视觉主动跟踪的非对称决斗机制。
IEEE Trans Pattern Anal Mach Intell. 2021 May;43(5):1467-1482. doi: 10.1109/TPAMI.2019.2952590. Epub 2021 Apr 1.
5
Jointly Feature Learning and Selection for Robust Tracking via a Gating Mechanism.通过门控机制进行鲁棒跟踪的联合特征学习与选择
PLoS One. 2016 Aug 30;11(8):e0161808. doi: 10.1371/journal.pone.0161808. eCollection 2016.
6
Hedging Deep Features for Visual Tracking.基于深度特征的视觉跟踪的套期保值。
IEEE Trans Pattern Anal Mach Intell. 2019 May;41(5):1116-1130. doi: 10.1109/TPAMI.2018.2828817. Epub 2018 Apr 20.
7
Real-Time 3D Facial Tracking via Cascaded Compositional Learning.基于级联组合学习的实时三维人脸跟踪。
IEEE Trans Image Process. 2021;30:3844-3857. doi: 10.1109/TIP.2021.3065819. Epub 2021 Mar 25.
8
Patch-based adaptive weighting with segmentation and scale (PAWSS) for visual tracking in surgical video.用于手术视频视觉跟踪的基于补丁的带分割和尺度的自适应加权(PAWSS)
Med Image Anal. 2019 Oct;57:120-135. doi: 10.1016/j.media.2019.07.002. Epub 2019 Jul 4.
9
Real-Time Human Detection for Aerial Captured Video Sequences via Deep Models.基于深度学习模型的航空捕获视频序列中的实时人体检测。
Comput Intell Neurosci. 2018 Feb 12;2018:1639561. doi: 10.1155/2018/1639561. eCollection 2018.
10
MBT3D: Deep learning based multi-object tracker for bumblebee 3D flight path estimation.基于深度学习的大黄蜂 3D 飞行轨迹估计的多目标跟踪器
PLoS One. 2023 Sep 22;18(9):e0291415. doi: 10.1371/journal.pone.0291415. eCollection 2023.

引用本文的文献

1
Mosaic: in-memory computing and routing for small-world spike-based neuromorphic systems.Mosaic:面向基于尖峰的小世界神经形态系统的内存计算与路由
Nat Commun. 2024 Jan 2;15(1):142. doi: 10.1038/s41467-023-44365-x.
2
When Autonomous Systems Meet Accuracy and Transferability through AI: A Survey.当自主系统通过人工智能实现准确性和可转移性时:一项综述。
Patterns (N Y). 2020 Jul 10;1(4):100050. doi: 10.1016/j.patter.2020.100050.
3
Perception-Action Coupling Target Tracking Control for a Snake Robot via Reinforcement Learning.基于强化学习的蛇形机器人感知-动作耦合目标跟踪控制
Front Neurorobot. 2020 Oct 20;14:591128. doi: 10.3389/fnbot.2020.591128. eCollection 2020.
4
Multiple-target tracking in human and machine vision.人类和机器视觉中的多目标跟踪。
PLoS Comput Biol. 2020 Apr 9;16(4):e1007698. doi: 10.1371/journal.pcbi.1007698. eCollection 2020 Apr.
5
Extreme Low-Light Image Enhancement for Surveillance Cameras Using Attention U-Net.基于注意力 U-Net 的监控摄像机极暗光图像增强。
Sensors (Basel). 2020 Jan 15;20(2):495. doi: 10.3390/s20020495.
6
Object Tracking for a Smart City Using IoT and Edge Computing.基于物联网和边缘计算的智慧城市目标跟踪
Sensors (Basel). 2019 Apr 28;19(9):1987. doi: 10.3390/s19091987.