• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于人体动作识别的稳健视频内容分析方案。

Robust video content analysis schemes for human action recognition.

机构信息

Faculty of Engineering and Technology, Multimedia University, Melaka, Malaysia.

出版信息

Sci Prog. 2021 Apr-Jun;104(2):368504211005480. doi: 10.1177/00368504211005480.

DOI:10.1177/00368504211005480
PMID:33913378
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10455027/
Abstract

INTRODUCTION

Action recognition is a challenging time series classification task that has received much attention in the recent past due to its importance in critical applications, such as surveillance, visual behavior study, topic discovery, security, and content retrieval.

OBJECTIVES

The main objective of the research is to develop a robust and high-performance human action recognition techniques. A combination of local and holistic feature extraction methods used through analyzing the most effective features to extract to reach the objective, followed by using simple and high-performance machine learning algorithms.

METHODS

This paper presents three robust action recognition techniques based on a series of image analysis methods to detect activities in different scenes. The general scheme architecture consists of shot boundary detection, shot frame rate re-sampling, and compact feature vector extraction. This process is achieved by emphasizing variations and extracting strong patterns in feature vectors before classification.

RESULTS

The proposed schemes are tested on datasets with cluttered backgrounds, low- or high-resolution videos, different viewpoints, and different camera motion conditions, namely, the Hollywood-2, KTH, UCF11 (YouTube actions), and Weizmann datasets. The proposed schemes resulted in highly accurate video analysis results compared to those of other works based on four widely used datasets. The First, Second, and Third Schemes provides recognition accuracies of 57.8%, 73.6%, and 52.0% on Hollywood2, 94.5%, 97.0%, and 59.3% on KTH, 94.5%, 95.6%, and 94.2% on UCF11, and 98.9%, 97.8% and 100% on Weizmann.

CONCLUSION

Each of the proposed schemes provides high recognition accuracy compared to other state-of-art methods. Especially, the Second Scheme as it gives excellent comparable results to other benchmarked approaches.

摘要

简介

动作识别是一项具有挑战性的时间序列分类任务,由于其在关键应用中的重要性,如监控、视觉行为研究、主题发现、安全和内容检索等,近年来受到了广泛关注。

目的

本研究的主要目的是开发一种强大且高性能的人体动作识别技术。通过分析最有效的特征提取方法,结合局部和整体特征提取方法,以达到目标,然后使用简单且高性能的机器学习算法。

方法

本文提出了三种基于一系列图像分析方法的鲁棒动作识别技术,用于检测不同场景中的活动。总体方案架构包括镜头边界检测、镜头帧率重采样和紧凑特征向量提取。通过强调特征向量中的变化并提取强模式来实现这一过程,然后再进行分类。

结果

所提出的方案在具有杂乱背景、低分辨率或高分辨率视频、不同视角和不同相机运动条件的数据集上进行了测试,即 Hollywood-2、KTH、UCF11(YouTube 动作)和 Weizmann 数据集。与基于四个广泛使用的数据集的其他工作相比,所提出的方案在视频分析结果方面具有很高的准确性。第一、第二和第三方案在 Hollywood2 上的识别准确率分别为 57.8%、73.6%和 52.0%,在 KTH 上的识别准确率分别为 94.5%、97.0%和 59.3%,在 UCF11 上的识别准确率分别为 94.5%、95.6%和 94.2%,在 Weizmann 上的识别准确率分别为 98.9%、97.8%和 100%。

结论

与其他最先进的方法相比,所提出的每个方案都提供了较高的识别准确率。特别是第二方案,因为它提供了与其他基准方法相当的结果。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/6e5a24b0e4a5/10.1177_00368504211005480-fig13.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/096aefa03cf0/10.1177_00368504211005480-fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/98f02361af89/10.1177_00368504211005480-fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/eff698eda18f/10.1177_00368504211005480-fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/0bd1dd059e4c/10.1177_00368504211005480-fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/04e4037e2dbf/10.1177_00368504211005480-fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/208836e5cc9c/10.1177_00368504211005480-fig6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/7f3decd7af03/10.1177_00368504211005480-fig7.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/06169ab11f3d/10.1177_00368504211005480-fig8.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/f2ecd72d273f/10.1177_00368504211005480-fig9.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/5e25d8482ef4/10.1177_00368504211005480-fig10.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/9838b3dd55d3/10.1177_00368504211005480-fig11.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/df5aebbd2154/10.1177_00368504211005480-fig12.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/6e5a24b0e4a5/10.1177_00368504211005480-fig13.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/096aefa03cf0/10.1177_00368504211005480-fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/98f02361af89/10.1177_00368504211005480-fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/eff698eda18f/10.1177_00368504211005480-fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/0bd1dd059e4c/10.1177_00368504211005480-fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/04e4037e2dbf/10.1177_00368504211005480-fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/208836e5cc9c/10.1177_00368504211005480-fig6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/7f3decd7af03/10.1177_00368504211005480-fig7.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/06169ab11f3d/10.1177_00368504211005480-fig8.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/f2ecd72d273f/10.1177_00368504211005480-fig9.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/5e25d8482ef4/10.1177_00368504211005480-fig10.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/9838b3dd55d3/10.1177_00368504211005480-fig11.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/df5aebbd2154/10.1177_00368504211005480-fig12.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3706/10455027/6e5a24b0e4a5/10.1177_00368504211005480-fig13.jpg

相似文献

1
Robust video content analysis schemes for human action recognition.用于人体动作识别的稳健视频内容分析方案。
Sci Prog. 2021 Apr-Jun;104(2):368504211005480. doi: 10.1177/00368504211005480.
2
Dynamic Spatio-Temporal Bag of Expressions (D-STBoE) Model for Human Action Recognition.用于人体动作识别的动态时空词袋(D-STBoE)模型。
Sensors (Basel). 2019 Jun 21;19(12):2790. doi: 10.3390/s19122790.
3
Learning Human Actions by Combining Global Dynamics and Local Appearance.通过组合全局动态和局部外观来学习人类动作。
IEEE Trans Pattern Anal Mach Intell. 2014 Dec;36(12):2466-82. doi: 10.1109/TPAMI.2014.2329301.
4
Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences.基于方向梯度直方图的动作视频序列中人体动作识别特征融合直方图
Sensors (Basel). 2020 Dec 18;20(24):7299. doi: 10.3390/s20247299.
5
Learning Spatio-Temporal Representations for Action Recognition: A Genetic Programming Approach.学习时空表示进行动作识别:一种遗传编程方法。
IEEE Trans Cybern. 2016 Jan;46(1):158-70. doi: 10.1109/TCYB.2015.2399172. Epub 2015 Feb 13.
6
Human action recognition based on HOIRM feature fusion and AP clustering BOW.基于 HOIRM 特征融合和 AP 聚类 BOW 的人体动作识别。
PLoS One. 2019 Jul 25;14(7):e0219910. doi: 10.1371/journal.pone.0219910. eCollection 2019.
7
Action recognition using mined hierarchical compound features.基于挖掘的层次化组合特征的动作识别。
IEEE Trans Pattern Anal Mach Intell. 2011 May;33(5):883-97. doi: 10.1109/TPAMI.2010.144.
8
Human Action Recognition Using Improved Salient Dense Trajectories.基于改进的显著密集轨迹的人体动作识别
Comput Intell Neurosci. 2016;2016:6750459. doi: 10.1155/2016/6750459. Epub 2016 May 17.
9
Discovering motion primitives for unsupervised grouping and one-shot learning of human actions, gestures, and expressions.发现运动基元,用于人类动作、手势和表情的无监督分组和一次性学习。
IEEE Trans Pattern Anal Mach Intell. 2013 Jul;35(7):1635-48. doi: 10.1109/TPAMI.2012.253.
10
Action Spotting and Recognition Based on a Spatiotemporal Orientation Analysis.基于时空方向分析的动作定位与识别。
IEEE Trans Pattern Anal Mach Intell. 2013 Mar;35(3):527-40. doi: 10.1109/TPAMI.2012.141.

本文引用的文献

1
View-invariant Deep Architecture for Human Action Recognition using Two-stream Motion and Shape Temporal Dynamics.使用双流运动和形状时间动态的用于人类动作识别的视图不变深度架构
IEEE Trans Image Process. 2020 Jan 15. doi: 10.1109/TIP.2020.2965299.
2
Hierarchical and Spatio-Temporal Sparse Representation for Human Action Recognition.层次化和时空稀疏表示的人体动作识别。
IEEE Trans Image Process. 2018 Apr;27(4):1748-1762. doi: 10.1109/TIP.2017.2788196.
3
Learning a Deep Model for Human Action Recognition from Novel Viewpoints.从新视角学习人类动作识别的深度模型。
IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):667-681. doi: 10.1109/TPAMI.2017.2691768. Epub 2017 Apr 6.
4
Hierarchical Clustering Multi-Task Learning for Joint Human Action Grouping and Recognition.层次聚类多任务学习用于联合人体动作分组和识别。
IEEE Trans Pattern Anal Mach Intell. 2017 Jan;39(1):102-114. doi: 10.1109/TPAMI.2016.2537337. Epub 2016 Mar 2.
5
Learning Spatio-Temporal Representations for Action Recognition: A Genetic Programming Approach.学习时空表示进行动作识别:一种遗传编程方法。
IEEE Trans Cybern. 2016 Jan;46(1):158-70. doi: 10.1109/TCYB.2015.2399172. Epub 2015 Feb 13.
6
Representing and retrieving video shots in human-centric brain imaging space.以人为主导的脑成像空间中的视频片段表示与检索。
IEEE Trans Image Process. 2013 Jul;22(7):2723-36. doi: 10.1109/TIP.2013.2256919. Epub 2013 Apr 4.
7
3D convolutional neural networks for human action recognition.三维卷积神经网络的人体动作识别。
IEEE Trans Pattern Anal Mach Intell. 2013 Jan;35(1):221-31. doi: 10.1109/TPAMI.2012.59.