• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于距离向量和多高视自适应网络的骨架动作识别。

Skeleton-Based Action Recognition Based on Distance Vector and Multihigh View Adaptive Networks.

机构信息

School of Computer Science, Hangzhou Dianzi University, Hangzhou 310000, China.

出版信息

Comput Intell Neurosci. 2021 Aug 18;2021:1507770. doi: 10.1155/2021/1507770. eCollection 2021.

DOI:10.1155/2021/1507770
PMID:34456989
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8390161/
Abstract

Skeleton-based human action recognition has attracted much attention in the field of computer vision. Most of the previous studies are based on fixed skeleton graphs so that only the local physical dependencies among joints can be captured, resulting in the omission of implicit joint correlations. In addition, under different views, the content of the same action is very different. In some views, keypoints will be blocked, which will cause recognition errors. In this paper, an action recognition method based on distance vector and multihigh view adaptive network (DV-MHNet) is proposed to address this challenging task. Among the mentioned techniques, the multihigh (MH) view adaptive networks are constructed to automatically determine the best observation view at different heights, obtain complete keypoints information of the current frame image, and enhance the robustness and generalization of the model to recognize actions at different heights. Then, the distance vector (DV) mechanism is introduced on this basis to establish the relative distance and relative orientation between different keypoints in the same frame and the same keypoints in different frame to obtain the global potential relationship of each keypoint, and finally by constructing the spatial temporal graph convolutional network to take into account the information in space and time, the characteristics of the action are learned. This paper has done the ablation study with traditional spatial temporal graph convolutional networks and with or without multihigh view adaptive networks, which reasonably proves the effectiveness of the model. The model is evaluated on two widely used action recognition benchmarks (NTU-RGB + D and PKU-MMD). Our method achieves better performance on both datasets.

摘要

基于骨架的人体动作识别在计算机视觉领域引起了广泛关注。之前的大多数研究都是基于固定的骨架图,因此只能捕捉到关节之间的局部物理依赖关系,从而忽略了隐含的关节相关性。此外,在不同视角下,同一动作的内容差异很大。在某些视角下,关键点会被遮挡,从而导致识别错误。本文提出了一种基于距离向量和多高视角自适应网络(DV-MHNet)的动作识别方法,以解决这一具有挑战性的任务。在提到的技术中,构建了多高(MH)视角自适应网络,以自动确定不同高度的最佳观察视角,获取当前帧图像的完整关键点信息,并增强模型在不同高度识别动作的鲁棒性和泛化能力。然后,在此基础上引入距离向量(DV)机制,建立同一帧内不同关键点和不同帧内同一关键点之间的相对距离和相对方向,获取每个关键点的全局潜在关系,最后通过构建时空图卷积网络来考虑空间和时间信息,学习动作的特征。本文通过与传统时空图卷积网络以及是否使用多高视角自适应网络进行了消融研究,合理地证明了模型的有效性。该模型在两个广泛使用的动作识别基准(NTU-RGB+D 和 PKU-MMD)上进行了评估。我们的方法在两个数据集上都取得了更好的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/370ed9c5edbf/CIN2021-1507770.008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/de8b0249dc0f/CIN2021-1507770.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/9db33c545de9/CIN2021-1507770.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/1e8ab7fc9488/CIN2021-1507770.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/0a201819f48a/CIN2021-1507770.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/1e8a937e62e6/CIN2021-1507770.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/2577f35d3117/CIN2021-1507770.006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/45fc7c4fc327/CIN2021-1507770.007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/370ed9c5edbf/CIN2021-1507770.008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/de8b0249dc0f/CIN2021-1507770.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/9db33c545de9/CIN2021-1507770.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/1e8ab7fc9488/CIN2021-1507770.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/0a201819f48a/CIN2021-1507770.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/1e8a937e62e6/CIN2021-1507770.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/2577f35d3117/CIN2021-1507770.006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/45fc7c4fc327/CIN2021-1507770.007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c40d/8390161/370ed9c5edbf/CIN2021-1507770.008.jpg

相似文献

1
Skeleton-Based Action Recognition Based on Distance Vector and Multihigh View Adaptive Networks.基于距离向量和多高视自适应网络的骨架动作识别。
Comput Intell Neurosci. 2021 Aug 18;2021:1507770. doi: 10.1155/2021/1507770. eCollection 2021.
2
Adaptive Attention Memory Graph Convolutional Networks for Skeleton-Based Action Recognition.用于基于骨架的动作识别的自适应注意力记忆图卷积网络
Sensors (Basel). 2021 Oct 12;21(20):6761. doi: 10.3390/s21206761.
3
Using Direct Acyclic Graphs to Enhance Skeleton-Based Action Recognition with a Linear-Map Convolution Neural Network.基于有向无环图的线性映射卷积神经网络在骨骼动作识别中的应用。
Sensors (Basel). 2021 Apr 29;21(9):3112. doi: 10.3390/s21093112.
4
Hypergraph Neural Network for Skeleton-Based Action Recognition.基于骨架的动作识别的超图神经网络。
IEEE Trans Image Process. 2021;30:2263-2275. doi: 10.1109/TIP.2021.3051495. Epub 2021 Jan 26.
5
GAS-GCN: Gated Action-Specific Graph Convolutional Networks for Skeleton-Based Action Recognition.GAS-GCN:基于骨骼的动作识别的门控动作特定图卷积网络。
Sensors (Basel). 2020 Jun 21;20(12):3499. doi: 10.3390/s20123499.
6
Multi-scale and attention enhanced graph convolution network for skeleton-based violence action recognition.用于基于骨架的暴力行为识别的多尺度注意力增强图卷积网络。
Front Neurorobot. 2022 Dec 15;16:1091361. doi: 10.3389/fnbot.2022.1091361. eCollection 2022.
7
Shallow Graph Convolutional Network for Skeleton-Based Action Recognition.基于骨架的动作识别的浅层图卷积网络。
Sensors (Basel). 2021 Jan 11;21(2):452. doi: 10.3390/s21020452.
8
Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition.用于基于骨架的动作识别的整体与部分自适应融合图卷积网络
Sensors (Basel). 2020 Dec 13;20(24):7149. doi: 10.3390/s20247149.
9
An efficient self-attention network for skeleton-based action recognition.基于骨架的动作识别的高效自注意力网络。
Sci Rep. 2022 Mar 8;12(1):4111. doi: 10.1038/s41598-022-08157-5.
10
Multi-Modality Adaptive Feature Fusion Graph Convolutional Network for Skeleton-Based Action Recognition.基于骨架的动作识别的多模态自适应特征融合图卷积网络。
Sensors (Basel). 2023 Jun 7;23(12):5414. doi: 10.3390/s23125414.

本文引用的文献

1
View Adaptive Neural Networks for High Performance Skeleton-Based Human Action Recognition.用于基于骨架的高性能人体动作识别的视图自适应神经网络。
IEEE Trans Pattern Anal Mach Intell. 2019 Aug;41(8):1963-1978. doi: 10.1109/TPAMI.2019.2896631. Epub 2019 Jan 31.
2
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.