• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于方向梯度直方图的动作视频序列中人体动作识别特征融合直方图

Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences.

作者信息

Patel Chirag I, Labana Dileep, Pandya Sharnil, Modi Kirit, Ghayvat Hemant, Awais Muhammad

机构信息

Computer Science & Engineering, Parul Institute of Technology, Parul University, Vadodara 391760, India.

Symbiosis Centre for Applied Artificial Intelligence and Symbiosis Institute of Technology, Symbiosis International (Deemed) University, Pune 412115, India.

出版信息

Sensors (Basel). 2020 Dec 18;20(24):7299. doi: 10.3390/s20247299.

DOI:10.3390/s20247299
PMID:33353248
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7766717/
Abstract

Human Action Recognition (HAR) is the classification of an action performed by a human. The goal of this study was to recognize human actions in action video sequences. We present a novel feature descriptor for HAR that involves multiple features and combining them using fusion technique. The major focus of the feature descriptor is to exploits the action dissimilarities. The key contribution of the proposed approach is to built robust features descriptor that can work for underlying video sequences and various classification models. To achieve the objective of the proposed work, HAR has been performed in the following manner. First, moving object detection and segmentation are performed from the background. The features are calculated using the histogram of oriented gradient (HOG) from a segmented moving object. To reduce the feature descriptor size, we take an averaging of the HOG features across non-overlapping video frames. For the frequency domain information we have calculated regional features from the Fourier hog. Moreover, we have also included the velocity and displacement of moving object. Finally, we use fusion technique to combine these features in the proposed work. After a feature descriptor is prepared, it is provided to the classifier. Here, we have used well-known classifiers such as artificial neural networks (ANNs), support vector machine (SVM), multiple kernel learning (MKL), Meta-cognitive Neural Network (McNN), and the late fusion methods. The main objective of the proposed approach is to prepare a robust feature descriptor and to show the diversity of our feature descriptor. Though we are using five different classifiers, our feature descriptor performs relatively well across the various classifiers. The proposed approach is performed and compared with the state-of-the-art methods for action recognition on two publicly available benchmark datasets (KTH and Weizmann) and for cross-validation on the UCF11 dataset, HMDB51 dataset, and UCF101 dataset. Results of the control experiments, such as a change in the SVM classifier and the effects of the second hidden layer in ANN, are also reported. The results demonstrate that the proposed method performs reasonably compared with the majority of existing state-of-the-art methods, including the convolutional neural network-based feature extractors.

摘要

人类行为识别(HAR)是对人类执行的动作进行分类。本研究的目标是在动作视频序列中识别人类行为。我们提出了一种用于HAR的新型特征描述符,它涉及多个特征并使用融合技术将它们组合起来。该特征描述符的主要重点是利用动作的差异。所提出方法的关键贡献在于构建能够适用于底层视频序列和各种分类模型的强大特征描述符。为了实现所提出工作的目标,HAR按以下方式进行。首先,从背景中进行运动物体检测和分割。使用从分割后的运动物体提取的方向梯度直方图(HOG)来计算特征。为了减小特征描述符的大小,我们对非重叠视频帧上的HOG特征进行平均。对于频域信息,我们从傅里叶HOG计算区域特征。此外,我们还纳入了运动物体的速度和位移。最后,我们在所提出的工作中使用融合技术来组合这些特征。在准备好特征描述符后,将其提供给分类器。在这里,我们使用了诸如人工神经网络(ANN)、支持向量机(SVM)、多核学习(MKL)、元认知神经网络(McNN)等知名分类器以及后期融合方法。所提出方法的主要目标是准备一个强大的特征描述符,并展示我们特征描述符的多样性。尽管我们使用了五种不同的分类器,但我们的特征描述符在各种分类器上的表现都相对较好。在所提出的方法在两个公开可用的基准数据集(KTH和魏茨曼)上进行动作识别,并在UCF11数据集、HMDB51数据集和UCF101数据集上进行交叉验证,并与当前最先进的方法进行比较。还报告了控制实验的结果,例如SVM分类器的变化以及ANN中第二个隐藏层的影响。结果表明,与大多数现有的当前最先进方法(包括基于卷积神经网络的特征提取器)相比,所提出的方法表现合理。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/4e033a7adc1f/sensors-20-07299-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/c322dbf37bcd/sensors-20-07299-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/9ad9cde7bccb/sensors-20-07299-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/197c8ac50a4e/sensors-20-07299-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/3e6df1f7aef7/sensors-20-07299-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/e0b9e4f354b6/sensors-20-07299-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/4dfc62404be6/sensors-20-07299-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/ffd41ceaffb9/sensors-20-07299-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/953b8271f64b/sensors-20-07299-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/b8c720bdf9df/sensors-20-07299-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/da875a8f1537/sensors-20-07299-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/581f720bb03c/sensors-20-07299-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/cb119a81feea/sensors-20-07299-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/a67f7d77a477/sensors-20-07299-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/4f944aaf7774/sensors-20-07299-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/32dea1b5c849/sensors-20-07299-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/683aa19f8ad5/sensors-20-07299-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/a7d1b2d33d02/sensors-20-07299-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/d26746c02853/sensors-20-07299-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/a10e9516819b/sensors-20-07299-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/4e033a7adc1f/sensors-20-07299-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/c322dbf37bcd/sensors-20-07299-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/9ad9cde7bccb/sensors-20-07299-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/197c8ac50a4e/sensors-20-07299-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/3e6df1f7aef7/sensors-20-07299-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/e0b9e4f354b6/sensors-20-07299-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/4dfc62404be6/sensors-20-07299-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/ffd41ceaffb9/sensors-20-07299-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/953b8271f64b/sensors-20-07299-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/b8c720bdf9df/sensors-20-07299-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/da875a8f1537/sensors-20-07299-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/581f720bb03c/sensors-20-07299-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/cb119a81feea/sensors-20-07299-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/a67f7d77a477/sensors-20-07299-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/4f944aaf7774/sensors-20-07299-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/32dea1b5c849/sensors-20-07299-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/683aa19f8ad5/sensors-20-07299-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/a7d1b2d33d02/sensors-20-07299-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/d26746c02853/sensors-20-07299-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/a10e9516819b/sensors-20-07299-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3a23/7766717/4e033a7adc1f/sensors-20-07299-g020.jpg

相似文献

1
Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences.基于方向梯度直方图的动作视频序列中人体动作识别特征融合直方图
Sensors (Basel). 2020 Dec 18;20(24):7299. doi: 10.3390/s20247299.
2
Robust video content analysis schemes for human action recognition.用于人体动作识别的稳健视频内容分析方案。
Sci Prog. 2021 Apr-Jun;104(2):368504211005480. doi: 10.1177/00368504211005480.
3
Automated classification of nasal polyps in endoscopy video-frames using handcrafted and CNN features.使用手工制作和 CNN 特征对内镜视频中的鼻息肉进行自动分类。
Comput Biol Med. 2022 Aug;147:105725. doi: 10.1016/j.compbiomed.2022.105725. Epub 2022 Jun 13.
4
Vision-Based HAR in UAV Videos Using Histograms and Deep Learning Techniques.基于视觉的无人机视频中的 HAR 研究:使用直方图和深度学习技术。
Sensors (Basel). 2023 Feb 25;23(5):2569. doi: 10.3390/s23052569.
5
Complex Human Action Recognition Using a Hierarchical Feature Reduction and Deep Learning-Based Method.基于层次特征约简和深度学习方法的复杂人类行为识别
SN Comput Sci. 2021;2(2):94. doi: 10.1007/s42979-021-00484-0. Epub 2021 Feb 13.
6
A multidimensional feature fusion network based on MGSE and TAAC for video-based human action recognition.一种基于MGSE和TAAC的多维度特征融合网络用于基于视频的人体动作识别。
Neural Netw. 2023 Nov;168:496-507. doi: 10.1016/j.neunet.2023.09.031. Epub 2023 Sep 22.
7
Enhancing Detection Quality Rate with a Combined HOG and CNN for Real-Time Multiple Object Tracking across Non-Overlapping Multiple Cameras.基于 HOG 和 CNN 的组合算法提高实时多目标跨非重叠多摄像机跟踪检测质量率。
Sensors (Basel). 2022 Mar 9;22(6):2123. doi: 10.3390/s22062123.
8
Feature Fusion of Deep Spatial Features and Handcrafted Spatiotemporal Features for Human Action Recognition.深度空间特征与手工制作时空特征的融合用于人体动作识别。
Sensors (Basel). 2019 Apr 2;19(7):1599. doi: 10.3390/s19071599.
9
Design and Development of an Imitation Detection System for Human Action Recognition Using Deep Learning.基于深度学习的人体动作识别模仿检测系统的设计与开发。
Sensors (Basel). 2023 Dec 18;23(24):9889. doi: 10.3390/s23249889.
10
Using a Selective Ensemble Support Vector Machine to Fuse Multimodal Features for Human Action Recognition.使用选择性集成支持向量机融合多模态特征进行人体动作识别。
Comput Intell Neurosci. 2022 Jan 10;2022:1877464. doi: 10.1155/2022/1877464. eCollection 2022.

引用本文的文献

1
A Comprehensive Methodological Survey of Human Activity Recognition Across Diverse Data Modalities.跨多种数据模态的人类活动识别综合方法学综述
Sensors (Basel). 2025 Jun 27;25(13):4028. doi: 10.3390/s25134028.
2
The analysis of motion recognition model for badminton player movements using machine learning.基于机器学习的羽毛球运动员动作运动识别模型分析
Sci Rep. 2025 May 30;15(1):19030. doi: 10.1038/s41598-025-02771-9.
3
The application of suitable sports games for junior high school students based on deep learning and artificial intelligence.

本文引用的文献

1
A Comprehensive Survey of Vision-Based Human Action Recognition Methods.基于视觉的人体动作识别方法综述。
Sensors (Basel). 2019 Feb 27;19(5):1005. doi: 10.3390/s19051005.
2
Long-Term Recurrent Convolutional Networks for Visual Recognition and Description.长期递归卷积网络的视觉识别与描述。
IEEE Trans Pattern Anal Mach Intell. 2017 Apr;39(4):677-691. doi: 10.1109/TPAMI.2016.2599174. Epub 2016 Sep 1.
3
Learning Spatio-Temporal Representations for Action Recognition: A Genetic Programming Approach.学习时空表示进行动作识别:一种遗传编程方法。
基于深度学习和人工智能的适合初中生的体育游戏应用
Sci Rep. 2025 May 16;15(1):17056. doi: 10.1038/s41598-025-01941-z.
4
A Review of Machine Learning and Deep Learning Methods for Person Detection, Tracking and Identification, and Face Recognition with Applications.用于人员检测、跟踪与识别以及人脸识别应用的机器学习和深度学习方法综述
Sensors (Basel). 2025 Feb 26;25(5):1410. doi: 10.3390/s25051410.
5
A Comprehensive Survey of Machine Learning Techniques and Models for Object Detection.目标检测的机器学习技术与模型综合调查
Sensors (Basel). 2025 Jan 2;25(1):214. doi: 10.3390/s25010214.
6
A New Multi-Branch Convolutional Neural Network and Feature Map Extraction Method for Traffic Congestion Detection.一种用于交通拥堵检测的新型多分支卷积神经网络及特征图提取方法
Sensors (Basel). 2024 Jul 1;24(13):4272. doi: 10.3390/s24134272.
7
SAR image matching based on rotation-invariant description.基于旋转不变描述的合成孔径雷达(SAR)图像匹配
Sci Rep. 2023 Sep 4;13(1):14510. doi: 10.1038/s41598-023-41592-6.
8
CNN-LSTM Model for Recognizing Video-Recorded Actions Performed in a Traditional Chinese Exercise.用于识别传统中国功法中视频记录动作的 CNN-LSTM 模型。
IEEE J Transl Eng Health Med. 2023 Jun 2;11:351-359. doi: 10.1109/JTEHM.2023.3282245. eCollection 2023.
9
Human Behavior Recognition via Hierarchical Patches Descriptor and Approximate Locality-Constrained Linear Coding.基于分层补丁描述符和近似局部约束线性编码的人类行为识别
Sensors (Basel). 2023 May 29;23(11):5179. doi: 10.3390/s23115179.
10
HFR-Video-Based Stereo Correspondence Using High Synchronous Short-Term Velocities.基于高同步短期速度的 HFR 视频立体匹配。
Sensors (Basel). 2023 Apr 26;23(9):4285. doi: 10.3390/s23094285.
IEEE Trans Cybern. 2016 Jan;46(1):158-70. doi: 10.1109/TCYB.2015.2399172. Epub 2015 Feb 13.
4
Actions as space-time shapes.作为时空形态的行动。
IEEE Trans Pattern Anal Mach Intell. 2007 Dec;29(12):2247-53. doi: 10.1109/TPAMI.2007.70711.