• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

关于时间动作建议生成的综合综述

A Comprehensive Review on Temporal-Action Proposal Generation.

作者信息

Sooksatra Sorn, Watcharapinchai Sitapa

机构信息

National Electronic and Computer Technology Center, National Science and Technology Development Agency, Pathum Thani 12120, Thailand.

出版信息

J Imaging. 2022 Jul 23;8(8):207. doi: 10.3390/jimaging8080207.

DOI:10.3390/jimaging8080207
PMID:35893085
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9394347/
Abstract

Temporal-action proposal generation (TAPG) is a well-known pre-processing of temporal-action localization and mainly affects localization performance on untrimmed videos. In recent years, there has been growing interest in proposal generation. Researchers have recently focused on anchor- and boundary-based methods for generating action proposals. The main purpose of this paper is to provide a comprehensive review of temporal-action proposal generation with network architectures and empirical results. The pre-processing step for input data is also discussed for network construction. The content of this paper was obtained from the research literature related to temporal-action proposal generation from 2012 to 2022 for performance evaluation and comparison. From several well-known databases, we used specific keywords to select 71 related studies according to their contributions and evaluation criteria. The contributions and methodologies are summarized and analyzed in a tabular form for each category. The result from state-of-the-art research was further analyzed to show its limitations and challenges for action proposal generation. TAPG performance in average recall ranges from 60% up to 78% in two TAPG benchmarks. In addition, several future potential research directions in this field are suggested based on the current limitations of the related studies.

摘要

时态动作建议生成(TAPG)是时态动作定位中一种广为人知的预处理方法,主要影响未修剪视频的定位性能。近年来,人们对建议生成的兴趣与日俱增。研究人员最近将重点放在基于锚点和基于边界的动作建议生成方法上。本文的主要目的是对具有网络架构和实证结果的时态动作建议生成进行全面综述。还讨论了网络构建中输入数据的预处理步骤。本文内容取自2012年至2022年与时态动作建议生成相关的研究文献,用于性能评估和比较。我们从几个知名数据库中,根据其贡献和评估标准,使用特定关键词选择了71项相关研究。对每一类别的贡献和方法以表格形式进行了总结和分析。对最先进研究的结果进行了进一步分析,以展示其在动作建议生成方面的局限性和挑战。在两个TAPG基准测试中,TAPG的平均召回率性能范围从60%到78%。此外,基于相关研究的当前局限性,提出了该领域未来几个潜在研究方向。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/7cfc23a83d81/jimaging-08-00207-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/2d6b8f9aa2a2/jimaging-08-00207-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/9781f3c122d3/jimaging-08-00207-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/b6ada2c51c22/jimaging-08-00207-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/7ed32b73d5cc/jimaging-08-00207-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/30dab81177e4/jimaging-08-00207-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/e79f78038eb6/jimaging-08-00207-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/3cc4f82b90f8/jimaging-08-00207-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/0ef5d92c69f3/jimaging-08-00207-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/f7f37bfdce8a/jimaging-08-00207-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/92d1ebc1fd73/jimaging-08-00207-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/377fb5c76a03/jimaging-08-00207-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/6bbe34b0f4ec/jimaging-08-00207-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/4e7190974491/jimaging-08-00207-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/bba830ffbb8f/jimaging-08-00207-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/e068d6bf21e8/jimaging-08-00207-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/7cfc23a83d81/jimaging-08-00207-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/2d6b8f9aa2a2/jimaging-08-00207-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/9781f3c122d3/jimaging-08-00207-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/b6ada2c51c22/jimaging-08-00207-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/7ed32b73d5cc/jimaging-08-00207-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/30dab81177e4/jimaging-08-00207-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/e79f78038eb6/jimaging-08-00207-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/3cc4f82b90f8/jimaging-08-00207-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/0ef5d92c69f3/jimaging-08-00207-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/f7f37bfdce8a/jimaging-08-00207-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/92d1ebc1fd73/jimaging-08-00207-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/377fb5c76a03/jimaging-08-00207-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/6bbe34b0f4ec/jimaging-08-00207-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/4e7190974491/jimaging-08-00207-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/bba830ffbb8f/jimaging-08-00207-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/e068d6bf21e8/jimaging-08-00207-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/9394347/7cfc23a83d81/jimaging-08-00207-g016.jpg

相似文献

1
A Comprehensive Review on Temporal-Action Proposal Generation.关于时间动作建议生成的综合综述
J Imaging. 2022 Jul 23;8(8):207. doi: 10.3390/jimaging8080207.
2
Spatio-Temporal Action Detection in Untrimmed Videos by Using Multimodal Features and Region Proposals.基于多模态特征和区域建议的非裁剪视频时空动作检测。
Sensors (Basel). 2019 Mar 3;19(5):1085. doi: 10.3390/s19051085.
3
Online action proposal generation using spatio-temporal attention network.基于时空注意力网络的在线动作建议生成。
Neural Netw. 2022 Sep;153:518-529. doi: 10.1016/j.neunet.2022.06.032. Epub 2022 Jun 30.
4
Multi-Level Content-Aware Boundary Detection for Temporal Action Proposal Generation.用于生成时间动作建议的多级内容感知边界检测
IEEE Trans Image Process. 2023;32:6090-6101. doi: 10.1109/TIP.2023.3328471. Epub 2023 Nov 8.
5
Weakly Supervised Temporal Action Localization Through Contrast Based Evaluation Networks.基于对比评估网络的弱监督时间动作定位
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):5886-5902. doi: 10.1109/TPAMI.2021.3078798. Epub 2022 Aug 4.
6
Deep Learning-Based Action Detection in Untrimmed Videos: A Survey.基于深度学习的未修剪视频动作检测:一项综述。
IEEE Trans Pattern Anal Mach Intell. 2023 Apr;45(4):4302-4320. doi: 10.1109/TPAMI.2022.3193611. Epub 2023 Mar 7.
7
RecapNet: Action Proposal Generation Mimicking Human Cognitive Process.RecapNet:模仿人类认知过程的动作提案生成。
IEEE Trans Cybern. 2021 Dec;51(12):6017-6028. doi: 10.1109/TCYB.2020.2965196. Epub 2021 Dec 22.
8
YoTube: Searching Action Proposal Via Recurrent and Static Regression Networks.YoTube:通过递归和静态回归网络进行搜索动作提案。
IEEE Trans Image Process. 2018 Jun;27(6):2609-2622. doi: 10.1109/TIP.2018.2806279.
9
Semantic and Temporal Contextual Correlation Learning for Weakly-Supervised Temporal Action Localization.用于弱监督时间动作定位的语义和时间上下文关联学习
IEEE Trans Pattern Anal Mach Intell. 2023 Oct;45(10):12427-12443. doi: 10.1109/TPAMI.2023.3287208. Epub 2023 Sep 5.
10
Adaptive Two-Stream Consensus Network for Weakly-Supervised Temporal Action Localization.自适应双流共识网络的弱监督时间动作定位。
IEEE Trans Pattern Anal Mach Intell. 2023 Apr;45(4):4136-4151. doi: 10.1109/TPAMI.2022.3189662. Epub 2023 Mar 7.

本文引用的文献

1
Temporal Segment Networks for Action Recognition in Videos.用于视频动作识别的时态片段网络
IEEE Trans Pattern Anal Mach Intell. 2019 Nov;41(11):2740-2755. doi: 10.1109/TPAMI.2018.2868668. Epub 2018 Sep 3.
2
Long-Term Temporal Convolutions for Action Recognition.长期时间卷积用于动作识别。
IEEE Trans Pattern Anal Mach Intell. 2018 Jun;40(6):1510-1517. doi: 10.1109/TPAMI.2017.2712608. Epub 2017 Jun 6.
3
Long-Term Recurrent Convolutional Networks for Visual Recognition and Description.长期递归卷积网络的视觉识别与描述。
IEEE Trans Pattern Anal Mach Intell. 2017 Apr;39(4):677-691. doi: 10.1109/TPAMI.2016.2599174. Epub 2016 Sep 1.
4
Human-level control through deep reinforcement learning.通过深度强化学习实现人类水平的控制。
Nature. 2015 Feb 26;518(7540):529-33. doi: 10.1038/nature14236.
5
3D convolutional neural networks for human action recognition.三维卷积神经网络的人体动作识别。
IEEE Trans Pattern Anal Mach Intell. 2013 Jan;35(1):221-31. doi: 10.1109/TPAMI.2012.59.