• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于自主系统控制与决策中动作识别的运动敏感网络。

Motion sensitive network for action recognition in control and decision-making of autonomous systems.

作者信息

Gu Jialiang, Yi Yang, Li Qiang

机构信息

Computer Science and Engineering, Sun Yat-sen University, Guangdong, China.

出版信息

Front Neurosci. 2024 Mar 25;18:1370024. doi: 10.3389/fnins.2024.1370024. eCollection 2024.

DOI:10.3389/fnins.2024.1370024
PMID:38591065
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11000707/
Abstract

Spatial-temporal modeling is crucial for action recognition in videos within the field of artificial intelligence. However, robustly extracting motion information remains a primary challenge due to temporal deformations of appearances and variations in motion frequencies between different actions. In order to address these issues, we propose an innovative and effective method called the Motion Sensitive Network (MSN), incorporating the theories of artificial neural networks and key concepts of autonomous system control and decision-making. Specifically, we employ an approach known as Spatial-Temporal Pyramid Motion Extraction (STP-ME) module, adjusting convolution kernel sizes and time intervals synchronously to gather motion information at different temporal scales, aligning with the learning and prediction characteristics of artificial neural networks. Additionally, we introduce a new module called Variable Scale Motion Excitation (DS-ME), utilizing a differential model to capture motion information in resonance with the flexibility of autonomous system control. Particularly, we employ a multi-scale deformable convolutional network to alter the motion scale of the target object before computing temporal differences across consecutive frames, providing theoretical support for the flexibility of autonomous systems. Temporal modeling is a crucial step in understanding environmental changes and actions within autonomous systems, and MSN, by integrating the advantages of Artificial Neural Networks (ANN) in this task, provides an effective framework for the future utilization of artificial neural networks in autonomous systems. We evaluate our proposed method on three challenging action recognition datasets (Kinetics-400, Something-Something V1, and Something-Something V2). The results indicate an improvement in accuracy ranging from 1.1% to 2.2% on the test set. When compared with state-of-the-art (SOTA) methods, the proposed approach achieves a maximum performance of 89.90%. In ablation experiments, the performance gain of this module also shows an increase ranging from 2% to 5.3%. The introduced Motion Sensitive Network (MSN) demonstrates significant potential in various challenging scenarios, providing an initial exploration into integrating artificial neural networks into the domain of autonomous systems.

摘要

时空建模对于人工智能领域视频中的动作识别至关重要。然而,由于外观的时间变形以及不同动作之间运动频率的变化,稳健地提取运动信息仍然是一个主要挑战。为了解决这些问题,我们提出了一种创新且有效的方法,称为运动敏感网络(MSN),它融合了人工神经网络的理论以及自主系统控制与决策的关键概念。具体而言,我们采用一种称为时空金字塔运动提取(STP - ME)模块的方法,同步调整卷积核大小和时间间隔,以在不同时间尺度上收集运动信息,这与人工神经网络的学习和预测特性相契合。此外,我们引入了一个名为可变尺度运动激励(DS - ME)的新模块,利用微分模型来捕捉与自主系统控制灵活性相共鸣的运动信息。特别地,我们采用多尺度可变形卷积网络在计算连续帧之间的时间差异之前改变目标对象的运动尺度,为自主系统的灵活性提供理论支持。时间建模是理解自主系统内环境变化和动作的关键步骤,而MSN通过在这项任务中整合人工神经网络(ANN)的优势,为未来在自主系统中利用人工神经网络提供了一个有效的框架。我们在三个具有挑战性的动作识别数据集(Kinetics - 400、Something - Something V1和Something - Something V2)上评估了我们提出的方法。结果表明,在测试集上准确率提高了1.1%至2.2%。与当前最先进(SOTA)方法相比,所提出的方法实现了89.90%的最高性能。在消融实验中,该模块的性能提升也显示出从2%到5.3%的增长。所引入的运动敏感网络(MSN)在各种具有挑战性的场景中展现出巨大潜力,为将人工神经网络集成到自主系统领域提供了初步探索。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/1de7389f25f8/fnins-18-1370024-g0012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/395e0e9a07f9/fnins-18-1370024-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/ede239fc8629/fnins-18-1370024-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/67c67e0efeaf/fnins-18-1370024-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/93688dcb9c84/fnins-18-1370024-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/3b8ce6c95b69/fnins-18-1370024-g0013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/0ad8a577c7b3/fnins-18-1370024-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/a6224bf93d1c/fnins-18-1370024-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/527ab05901ae/fnins-18-1370024-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/ee865c84c792/fnins-18-1370024-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/06c2c721cc29/fnins-18-1370024-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/d8b4dad64a90/fnins-18-1370024-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/7f083c8d118a/fnins-18-1370024-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/1de7389f25f8/fnins-18-1370024-g0012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/395e0e9a07f9/fnins-18-1370024-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/ede239fc8629/fnins-18-1370024-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/67c67e0efeaf/fnins-18-1370024-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/93688dcb9c84/fnins-18-1370024-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/3b8ce6c95b69/fnins-18-1370024-g0013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/0ad8a577c7b3/fnins-18-1370024-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/a6224bf93d1c/fnins-18-1370024-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/527ab05901ae/fnins-18-1370024-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/ee865c84c792/fnins-18-1370024-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/06c2c721cc29/fnins-18-1370024-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/d8b4dad64a90/fnins-18-1370024-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/7f083c8d118a/fnins-18-1370024-g0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/51fb/11000707/1de7389f25f8/fnins-18-1370024-g0012.jpg

相似文献

1
Motion sensitive network for action recognition in control and decision-making of autonomous systems.用于自主系统控制与决策中动作识别的运动敏感网络。
Front Neurosci. 2024 Mar 25;18:1370024. doi: 10.3389/fnins.2024.1370024. eCollection 2024.
2
A Spatio-Temporal Motion Network for Action Recognition Based on Spatial Attention.基于空间注意力的用于动作识别的时空运动网络
Entropy (Basel). 2022 Mar 4;24(3):368. doi: 10.3390/e24030368.
3
Leveraging spatial residual attention and temporal Markov networks for video action understanding.利用空间残差注意力和时间马尔可夫网络进行视频动作理解。
Neural Netw. 2024 Jan;169:378-387. doi: 10.1016/j.neunet.2023.10.047. Epub 2023 Oct 31.
4
MEST: An Action Recognition Network with Motion Encoder and Spatio-Temporal Module.MEST:一种具有运动编码器和时空模块的动作识别网络。
Sensors (Basel). 2022 Sep 1;22(17):6595. doi: 10.3390/s22176595.
5
Spatial-Temporal Pyramid Graph Reasoning for Action Recognition.用于动作识别的时空金字塔图推理
IEEE Trans Image Process. 2022;31:5484-5497. doi: 10.1109/TIP.2022.3196175. Epub 2022 Aug 22.
6
Non-Local Temporal Difference Network for Temporal Action Detection.非局部时域差分网络的时域动作检测。
Sensors (Basel). 2022 Nov 1;22(21):8396. doi: 10.3390/s22218396.
7
AMS-Net: Modeling Adaptive Multi-Granularity Spatio-Temporal Cues for Video Action Recognition.AMS-Net:用于视频动作识别的自适应多粒度时空线索建模
IEEE Trans Neural Netw Learn Syst. 2024 Dec;35(12):18731-18745. doi: 10.1109/TNNLS.2023.3321141. Epub 2024 Dec 2.
8
Learning SpatioTemporal and Motion Features in a Unified 2D Network for Action Recognition.在统一二维网络中学习用于动作识别的时空和运动特征
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3347-3362. doi: 10.1109/TPAMI.2022.3173658. Epub 2023 Feb 3.
9
Multiple Traffic Target Tracking with Spatial-Temporal Affinity Network.基于时空关联网络的多目标跟踪
Comput Intell Neurosci. 2022 May 23;2022:9693767. doi: 10.1155/2022/9693767. eCollection 2022.
10
Human action recognition method based on Motion Excitation and Temporal Aggregation module.基于运动激励和时间聚合模块的人体动作识别方法
Heliyon. 2022 Nov 4;8(11):e11401. doi: 10.1016/j.heliyon.2022.e11401. eCollection 2022 Nov.

引用本文的文献

1
LLaVA-Pose: Keypoint-Integrated Instruction Tuning for Human Pose and Action Understanding.LLaVA-Pose:用于人体姿态与动作理解的关键点集成指令微调
Sensors (Basel). 2025 Aug 21;25(16):5213. doi: 10.3390/s25165213.
2
Artificial Intelligence and Its Revolutionary Role in Physical and Mental Rehabilitation: A Review of Recent Advancements.人工智能及其在身心康复中的变革性作用:近期进展综述
Biomed Res Int. 2024 Dec 17;2024:9554590. doi: 10.1155/bmri/9554590. eCollection 2024.

本文引用的文献

1
ProtoHAR: Prototype Guided Personalized Federated Learning for Human Activity Recognition.ProtoHAR:用于人体活动识别的原型指导个性化联邦学习。
IEEE J Biomed Health Inform. 2023 Aug;27(8):3900-3911. doi: 10.1109/JBHI.2023.3275438. Epub 2023 Aug 7.
2
Strategy to combine two functional components: Efficient nano material development for iodine immobilization.策略结合两种功能组件:碘固定用高效纳米材料的开发。
Chemosphere. 2022 Dec;309(Pt 1):136477. doi: 10.1016/j.chemosphere.2022.136477. Epub 2022 Sep 23.
3
Human Action Recognition From Various Data Modalities: A Review.
基于多种数据模态的人类行为识别综述
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3200-3225. doi: 10.1109/TPAMI.2022.3183112. Epub 2023 Feb 3.
4
Spectroscopically clean Au nanoparticles for catalytic decomposition of hydrogen peroxide.用于过氧化氢催化分解的光谱纯净金纳米颗粒。
Sci Rep. 2021 May 6;11(1):9709. doi: 10.1038/s41598-021-89235-y.