• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

利用 RBM-NN 的新颖参数初始化技术进行人体动作识别。

A Novel Parameter Initialization Technique Using RBM-NN for Human Action Recognition.

机构信息

DCSE, CEG-Anna University, Guindy, Chennai, India.

RCC, CEG-Anna University, Guindy, Chennai, India.

出版信息

Comput Intell Neurosci. 2020 Sep 10;2020:8852404. doi: 10.1155/2020/8852404. eCollection 2020.

DOI:10.1155/2020/8852404
PMID:32963513
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7501562/
Abstract

Human action recognition is a trending topic in the field of computer vision and its allied fields. The goal of human action recognition is to identify any human action that takes place in an image or a video dataset. For instance, the actions include walking, running, jumping, throwing, and much more. Existing human action recognition techniques have their own set of limitations when it concerns model accuracy and flexibility. To overcome these limitations, deep learning technologies were implemented. In the deep learning approach, a model learns by itself to improve its recognition accuracy and avoids problems such as gradient eruption, overfitting, and underfitting. In this paper, we propose a novel parameter initialization technique using the Maxout activation function. Firstly, human action is detected and tracked from the video dataset to learn the spatial-temporal features. Secondly, the extracted feature descriptors are trained using the RBM-NN. Thirdly, the local features are encoded into global features using an integrated forward and backward propagation process via RBM-NN. Finally, an SVM classifier recognizes the human actions in the video dataset. The experimental analysis performed on various benchmark datasets showed an improved recognition rate when compared to other state-of-the-art learning models.

摘要

人体动作识别是计算机视觉及其相关领域的一个热门话题。人体动作识别的目标是识别图像或视频数据集中发生的任何人体动作。例如,这些动作包括行走、跑步、跳跃、投掷等等。现有的人体动作识别技术在模型准确性和灵活性方面都存在自身的局限性。为了克服这些局限性,人们采用了深度学习技术。在深度学习方法中,模型通过自我学习来提高识别准确性,并避免了梯度爆炸、过拟合和欠拟合等问题。在本文中,我们提出了一种新的参数初始化技术,使用 Maxout 激活函数。首先,从视频数据集中检测和跟踪人体动作,以学习时空特征。其次,使用 RBM-NN 对提取的特征描述符进行训练。然后,通过 RBM-NN 的正向和反向传播过程,将局部特征编码为全局特征。最后,使用 SVM 分类器识别视频数据集中的人体动作。在各种基准数据集上进行的实验分析表明,与其他最先进的学习模型相比,识别率得到了提高。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/22f8a253072b/CIN2020-8852404.025.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/6cfbf339f545/CIN2020-8852404.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/e07fd48cb9b2/CIN2020-8852404.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/fa1d07ff8680/CIN2020-8852404.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/59538bfa3f9c/CIN2020-8852404.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/ff935198a31a/CIN2020-8852404.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/c85cd054cf55/CIN2020-8852404.006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/326fe5df3818/CIN2020-8852404.007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/13b1718366f6/CIN2020-8852404.008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/3e3415bbd94d/CIN2020-8852404.009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/f5569c8ce831/CIN2020-8852404.010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/d5f8674899b6/CIN2020-8852404.011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/5074dc05ba98/CIN2020-8852404.012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/d6b432945fa5/CIN2020-8852404.013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/a3d8142118bb/CIN2020-8852404.014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/bf1a6475e23a/CIN2020-8852404.015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/8d317c3098c3/CIN2020-8852404.016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/3587542c748e/CIN2020-8852404.017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/5eaa3178de47/CIN2020-8852404.018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/b16e41ca519f/CIN2020-8852404.019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/31c9797559d9/CIN2020-8852404.020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/dca0ac680387/CIN2020-8852404.021.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/b4b3d09d3de2/CIN2020-8852404.022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/c2b33f22755f/CIN2020-8852404.023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/8493a3936fc9/CIN2020-8852404.024.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/22f8a253072b/CIN2020-8852404.025.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/6cfbf339f545/CIN2020-8852404.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/e07fd48cb9b2/CIN2020-8852404.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/fa1d07ff8680/CIN2020-8852404.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/59538bfa3f9c/CIN2020-8852404.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/ff935198a31a/CIN2020-8852404.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/c85cd054cf55/CIN2020-8852404.006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/326fe5df3818/CIN2020-8852404.007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/13b1718366f6/CIN2020-8852404.008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/3e3415bbd94d/CIN2020-8852404.009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/f5569c8ce831/CIN2020-8852404.010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/d5f8674899b6/CIN2020-8852404.011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/5074dc05ba98/CIN2020-8852404.012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/d6b432945fa5/CIN2020-8852404.013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/a3d8142118bb/CIN2020-8852404.014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/bf1a6475e23a/CIN2020-8852404.015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/8d317c3098c3/CIN2020-8852404.016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/3587542c748e/CIN2020-8852404.017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/5eaa3178de47/CIN2020-8852404.018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/b16e41ca519f/CIN2020-8852404.019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/31c9797559d9/CIN2020-8852404.020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/dca0ac680387/CIN2020-8852404.021.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/b4b3d09d3de2/CIN2020-8852404.022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/c2b33f22755f/CIN2020-8852404.023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/8493a3936fc9/CIN2020-8852404.024.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ac24/7501562/22f8a253072b/CIN2020-8852404.025.jpg

相似文献

1
A Novel Parameter Initialization Technique Using RBM-NN for Human Action Recognition.利用 RBM-NN 的新颖参数初始化技术进行人体动作识别。
Comput Intell Neurosci. 2020 Sep 10;2020:8852404. doi: 10.1155/2020/8852404. eCollection 2020.
2
Human Action Recognition Using Improved Salient Dense Trajectories.基于改进的显著密集轨迹的人体动作识别
Comput Intell Neurosci. 2016;2016:6750459. doi: 10.1155/2016/6750459. Epub 2016 May 17.
3
Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences.基于方向梯度直方图的动作视频序列中人体动作识别特征融合直方图
Sensors (Basel). 2020 Dec 18;20(24):7299. doi: 10.3390/s20247299.
4
Robust video content analysis schemes for human action recognition.用于人体动作识别的稳健视频内容分析方案。
Sci Prog. 2021 Apr-Jun;104(2):368504211005480. doi: 10.1177/00368504211005480.
5
C-MHAD: Continuous Multimodal Human Action Dataset of Simultaneous Video and Inertial Sensing.C-MHAD:同时视频和惯性感知的连续多模态人体动作数据集。
Sensors (Basel). 2020 May 20;20(10):2905. doi: 10.3390/s20102905.
6
Learning a Deep Model for Human Action Recognition from Novel Viewpoints.从新视角学习人类动作识别的深度模型。
IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):667-681. doi: 10.1109/TPAMI.2017.2691768. Epub 2017 Apr 6.
7
Feature Fusion of Deep Spatial Features and Handcrafted Spatiotemporal Features for Human Action Recognition.深度空间特征与手工制作时空特征的融合用于人体动作识别。
Sensors (Basel). 2019 Apr 2;19(7):1599. doi: 10.3390/s19071599.
8
Basketball technique action recognition using 3D convolutional neural networks.基于 3D 卷积神经网络的篮球技术动作识别
Sci Rep. 2024 Jun 7;14(1):13156. doi: 10.1038/s41598-024-63621-8.
9
Dynamic Spatio-Temporal Bag of Expressions (D-STBoE) Model for Human Action Recognition.用于人体动作识别的动态时空词袋(D-STBoE)模型。
Sensors (Basel). 2019 Jun 21;19(12):2790. doi: 10.3390/s19122790.
10
Desktop Action Recognition From First-Person Point-of-View.基于第一人称视角的桌面行为识别。
IEEE Trans Cybern. 2019 May;49(5):1616-1628. doi: 10.1109/TCYB.2018.2806381. Epub 2018 Feb 27.

引用本文的文献

1
Wearable Sensor-Based Human Activity Recognition in the Smart Healthcare System.基于可穿戴传感器的智能医疗保健系统中的人体活动识别。
Comput Intell Neurosci. 2022 Feb 24;2022:1391906. doi: 10.1155/2022/1391906. eCollection 2022.

本文引用的文献

1
Global-Local Temporal Saliency Action Prediction.全局-局部时空显著性动作预测。
IEEE Trans Image Process. 2018 May;27(5):2272-2285. doi: 10.1109/TIP.2017.2751145. Epub 2017 Sep 11.
2
Long-Term Recurrent Convolutional Networks for Visual Recognition and Description.长期递归卷积网络的视觉识别与描述。
IEEE Trans Pattern Anal Mach Intell. 2017 Apr;39(4):677-691. doi: 10.1109/TPAMI.2016.2599174. Epub 2016 Sep 1.
3
Action Recognition Using Rate-Invariant Analysis of Skeletal Shape Trajectories.基于骨骼形状轨迹的不变率分析的动作识别。
IEEE Trans Pattern Anal Mach Intell. 2016 Jan;38(1):1-13. doi: 10.1109/tpami.2015.2439257.
4
Max-Margin Action Prediction Machine.最大间隔动作预测机。
IEEE Trans Pattern Anal Mach Intell. 2016 Sep;38(9):1844-58. doi: 10.1109/TPAMI.2015.2491928. Epub 2015 Oct 16.
5
Learning Human Actions by Combining Global Dynamics and Local Appearance.通过组合全局动态和局部外观来学习人类动作。
IEEE Trans Pattern Anal Mach Intell. 2014 Dec;36(12):2466-82. doi: 10.1109/TPAMI.2014.2329301.
6
Interactive Phrases: Semantic Descriptions for Human Interaction Recognition.互动短语:用于人类交互识别的语义描述。
IEEE Trans Pattern Anal Mach Intell. 2014 Sep;36(9):1775-88. doi: 10.1109/TPAMI.2014.2303090.
7
Modeling Geometric-Temporal Context With Directional Pyramid Co-Occurrence for Action Recognition.基于方向金字塔共现的时空上下文建模方法及其在动作识别中的应用
IEEE Trans Image Process. 2014 Feb;23(2):658-72. doi: 10.1109/TIP.2013.2291319.
8
Hessian-regularized co-training for social activity recognition.用于社交活动识别的黑森正则化协同训练
PLoS One. 2014 Sep 26;9(9):e108474. doi: 10.1371/journal.pone.0108474. eCollection 2014.
9
Spatio-temporal Laplacian pyramid coding for action recognition.基于时空拉普拉斯金字塔的动作识别。
IEEE Trans Cybern. 2014 Jun;44(6):817-27. doi: 10.1109/TCYB.2013.2273174. Epub 2013 Jul 31.
10
Representation learning: a review and new perspectives.表示学习:综述与新视角。
IEEE Trans Pattern Anal Mach Intell. 2013 Aug;35(8):1798-828. doi: 10.1109/TPAMI.2013.50.