• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于弱人类活动定位的深度强化学习

Deep reinforcement learning for weak human activity localization.

作者信息

Xu Wanru, Miao Zhenjiang, Yu Jian, Ji Qiang

出版信息

IEEE Trans Image Process. 2019 Sep 26. doi: 10.1109/TIP.2019.2942814.

DOI:10.1109/TIP.2019.2942814
PMID:31567088
Abstract

Human activity localization aims at recognizing contents and detecting locations of activities in video sequences. With an increasing number of untrimmed video data, traditional activity localization methods always suffer from two major limitations. First, detailed annotations are needed in most existing methods, i.e., bounding-box annotations in every frame, which are both expensive and time consuming. Second, the search space is too large for 3D activity localization, which requires generating a large number of proposals. In this paper, we propose a unified deep Q-network with weak reward and weak loss (DWRLQN) to address the two problems. Certain weak knowledge and weak constraints involving the temporal dynamics of human activity are incorporated into a deep reinforcement learning framework under sparse spatial supervision, where we assume that only a portion of frames are annotated in each video sequence. Experiments on UCF-Sports, UCF-101 and sub-JHMDB demonstrate that our proposed model achieves promising performance by only utilizing a very small number of proposals. More importantly, our DWRLQN trained with partial annotations and weak information even outperforms fully supervised methods.

摘要

人类活动定位旨在识别视频序列中的活动内容并检测其位置。随着未修剪视频数据数量的不断增加,传统的活动定位方法一直存在两个主要局限性。首先,大多数现有方法都需要详细的注释,即在每一帧中进行边界框注释,这既昂贵又耗时。其次,对于三维活动定位而言,搜索空间太大,这需要生成大量的提议。在本文中,我们提出了一种具有弱奖励和弱损失的统一深度Q网络(DWRLQN)来解决这两个问题。在稀疏空间监督下,将涉及人类活动时间动态的某些弱知识和弱约束纳入深度强化学习框架,我们假设在每个视频序列中只有一部分帧被注释。在UCF-Sports、UCF-101和子JHMDB上的实验表明,我们提出的模型仅通过使用非常少量的提议就取得了有前景的性能。更重要的是,我们用部分注释和弱信息训练的DWRLQN甚至优于完全监督的方法。

相似文献

1
Deep reinforcement learning for weak human activity localization.用于弱人类活动定位的深度强化学习
IEEE Trans Image Process. 2019 Sep 26. doi: 10.1109/TIP.2019.2942814.
2
YoTube: Searching Action Proposal Via Recurrent and Static Regression Networks.YoTube:通过递归和静态回归网络进行搜索动作提案。
IEEE Trans Image Process. 2018 Jun;27(6):2609-2622. doi: 10.1109/TIP.2018.2806279.
3
Deep Graph Metric Learning for Weakly Supervised Person Re-Identification.深度图度量学习在弱监督行人再识别中的应用。
IEEE Trans Pattern Anal Mach Intell. 2022 Oct;44(10):6074-6093. doi: 10.1109/TPAMI.2021.3084613. Epub 2022 Sep 14.
4
Single-Frame Supervision for Spatio-Temporal Video Grounding.用于时空视频定位的单帧监督
IEEE Trans Pattern Anal Mach Intell. 2025 Jul;47(7):5177-5191. doi: 10.1109/TPAMI.2024.3415087.
5
Progressive Frame-Proposal Mining for Weakly Supervised Video Object Detection.用于弱监督视频目标检测的渐进式帧提议挖掘
IEEE Trans Image Process. 2024;33:1560-1573. doi: 10.1109/TIP.2024.3364536. Epub 2024 Feb 27.
6
Combining weakly and strongly supervised learning improves strong supervision in Gleason pattern classification.弱监督和强监督学习的结合提高了 Gleason 模式分类中的强监督。
BMC Med Imaging. 2021 May 8;21(1):77. doi: 10.1186/s12880-021-00609-0.
7
Adaptive Two-Stream Consensus Network for Weakly-Supervised Temporal Action Localization.自适应双流共识网络的弱监督时间动作定位。
IEEE Trans Pattern Anal Mach Intell. 2023 Apr;45(4):4136-4151. doi: 10.1109/TPAMI.2022.3189662. Epub 2023 Mar 7.
8
Weakly Supervised Semantic Segmentation via Box-Driven Masking and Filling Rate Shifting.通过框驱动掩蔽和填充率转移实现弱监督语义分割
IEEE Trans Pattern Anal Mach Intell. 2023 Dec;45(12):15996-16012. doi: 10.1109/TPAMI.2023.3301302. Epub 2023 Nov 3.
9
DANet: Semi-supervised differentiated auxiliaries guided network for video action recognition.DANet:用于视频动作识别的半监督差异化辅助引导网络。
Neural Netw. 2023 Jan;158:121-131. doi: 10.1016/j.neunet.2022.11.009. Epub 2022 Nov 17.
10
SPFTN: A Joint Learning Framework for Localizing and Segmenting Objects in Weakly Labeled Videos.SPFTN:一种用于在弱标注视频中定位和分割对象的联合学习框架。
IEEE Trans Pattern Anal Mach Intell. 2020 Feb;42(2):475-489. doi: 10.1109/TPAMI.2018.2881114. Epub 2018 Nov 13.

引用本文的文献

1
Human Activity Recognition Data Analysis: History, Evolutions, and New Trends.人体活动识别数据分析:历史、发展与新趋势。
Sensors (Basel). 2022 Apr 29;22(9):3401. doi: 10.3390/s22093401.