• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于基于骨架的动作识别的语义辅助训练图卷积网络

Semantics-Assisted Training Graph Convolution Network for Skeleton-Based Action Recognition.

作者信息

Hu Huangshui, Cao Yu, Fang Yue, Meng Zhiqiang

机构信息

College of Computer Science and Engineering, Changchun University of Technology, Changchun 130012, China.

出版信息

Sensors (Basel). 2025 Mar 15;25(6):1841. doi: 10.3390/s25061841.

DOI:10.3390/s25061841
PMID:40292967
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11946682/
Abstract

The skeleton-based action recognition networks often focus on extracting features such as joints from samples, while neglecting the semantic relationships inherent in actions, which also contain valuable information. To address the lack of utilization of semantic information, this paper proposes a semantics-assisted training graph convolution network (SAT-GCN). By dividing the features outputted by the skeleton encoder into four parts and contrasting them with the text features generated by the text encoder, the obtained contrastive loss is used to guide the overall network training. This approach effectively improves recognition accuracy while reducing the number of model parameters. In addition, angle features are incorporated into the skeleton model input to aid in classifying similar actions. Finally, a multi-feature skeleton encoder is designed to separately extract features such as joints, bones, and angles. These extracted features are then integrated through feature fusion. The fused features are then passed through three graph convolution blocks before being fed into fully connected layers for classification. Extensive experiments were conducted on three large-scale datasets, NTU RGB + D 60, NTU RGB + D 120, and NW-UCLA to validate the performance of the proposed model. The results show that the SAT-GCN outperforms others in terms of both accuracy and number of parameters.

摘要

基于骨架的动作识别网络通常专注于从样本中提取诸如关节等特征,却忽略了动作中固有的语义关系,而这些语义关系也包含有价值的信息。为了解决语义信息利用不足的问题,本文提出了一种语义辅助训练图卷积网络(SAT-GCN)。通过将骨架编码器输出的特征分为四个部分,并将它们与文本编码器生成的文本特征进行对比,所获得的对比损失用于指导整个网络的训练。这种方法在减少模型参数数量的同时有效提高了识别准确率。此外,将角度特征纳入骨架模型输入以辅助对相似动作进行分类。最后,设计了一个多特征骨架编码器来分别提取诸如关节、骨骼和角度等特征。然后通过特征融合将这些提取的特征进行整合。融合后的特征在被送入全连接层进行分类之前先经过三个图卷积块。在三个大规模数据集NTU RGB + D 60、NTU RGB + D 120和NW-UCLA上进行了大量实验,以验证所提出模型的性能。结果表明,SAT-GCN在准确率和参数数量方面均优于其他模型。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/9db5e2a69da1/sensors-25-01841-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/94b8a1a88eae/sensors-25-01841-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/e3d71ecac8c3/sensors-25-01841-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/4d11ca949ae9/sensors-25-01841-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/9c73f6f36543/sensors-25-01841-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/c871920a4616/sensors-25-01841-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/8775d8ab992d/sensors-25-01841-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/9db5e2a69da1/sensors-25-01841-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/94b8a1a88eae/sensors-25-01841-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/e3d71ecac8c3/sensors-25-01841-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/4d11ca949ae9/sensors-25-01841-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/9c73f6f36543/sensors-25-01841-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/c871920a4616/sensors-25-01841-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/8775d8ab992d/sensors-25-01841-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/edf6/11946682/9db5e2a69da1/sensors-25-01841-g007.jpg

相似文献

1
Semantics-Assisted Training Graph Convolution Network for Skeleton-Based Action Recognition.用于基于骨架的动作识别的语义辅助训练图卷积网络
Sensors (Basel). 2025 Mar 15;25(6):1841. doi: 10.3390/s25061841.
2
Multi-scale and attention enhanced graph convolution network for skeleton-based violence action recognition.用于基于骨架的暴力行为识别的多尺度注意力增强图卷积网络。
Front Neurorobot. 2022 Dec 15;16:1091361. doi: 10.3389/fnbot.2022.1091361. eCollection 2022.
3
Multi-Modality Adaptive Feature Fusion Graph Convolutional Network for Skeleton-Based Action Recognition.基于骨架的动作识别的多模态自适应特征融合图卷积网络。
Sensors (Basel). 2023 Jun 7;23(12):5414. doi: 10.3390/s23125414.
4
Graph Diffusion Convolutional Network for Skeleton Based Semantic Recognition of Two-Person Actions.基于骨架的两人动作语义识别的图扩散卷积网络。
IEEE Trans Pattern Anal Mach Intell. 2023 Jul;45(7):8477-8493. doi: 10.1109/TPAMI.2023.3238411. Epub 2023 Jun 5.
5
TFC-GCN: Lightweight Temporal Feature Cross-Extraction Graph Convolutional Network for Skeleton-Based Action Recognition.TFC-GCN:基于骨架的动作识别的轻量级时间特征交叉提取图卷积网络。
Sensors (Basel). 2023 Jun 15;23(12):5593. doi: 10.3390/s23125593.
6
Whole and Part Adaptive Fusion Graph Convolutional Networks for Skeleton-Based Action Recognition.用于基于骨架的动作识别的整体与部分自适应融合图卷积网络
Sensors (Basel). 2020 Dec 13;20(24):7149. doi: 10.3390/s20247149.
7
Two-stream spatio-temporal GCN-transformer networks for skeleton-based action recognition.用于基于骨架的动作识别的双流时空GCN-Transformer网络
Sci Rep. 2025 Feb 10;15(1):4982. doi: 10.1038/s41598-025-87752-8.
8
Feedback Graph Convolutional Network for Skeleton-Based Action Recognition.用于基于骨架的动作识别的反馈图卷积网络
IEEE Trans Image Process. 2022;31:164-175. doi: 10.1109/TIP.2021.3129117. Epub 2021 Dec 2.
9
SelfGCN: Graph Convolution Network With Self-Attention for Skeleton-Based Action Recognition.SelfGCN:基于自注意力机制的用于基于骨架的动作识别的图卷积网络
IEEE Trans Image Process. 2024;33:4391-4403. doi: 10.1109/TIP.2024.3433581. Epub 2024 Aug 5.
10
Shallow Graph Convolutional Network for Skeleton-Based Action Recognition.基于骨架的动作识别的浅层图卷积网络。
Sensors (Basel). 2021 Jan 11;21(2):452. doi: 10.3390/s21020452.

本文引用的文献

1
Laparoscopic Image-Based Critical Action Recognition and Anticipation With Explainable Features.基于腹腔镜图像的关键动作识别与可解释特征预测
IEEE J Biomed Health Inform. 2023 Nov;27(11):5393-5404. doi: 10.1109/JBHI.2023.3306818. Epub 2023 Nov 7.
2
Visual Gait Analysis Based on UE4.基于 UE4 的视觉步态分析。
Sensors (Basel). 2023 Jun 9;23(12):5463. doi: 10.3390/s23125463.
3
Constructing Stronger and Faster Baselines for Skeleton-Based Action Recognition.为基于骨骼的动作识别构建更强更快的基线
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):1474-1488. doi: 10.1109/TPAMI.2022.3157033. Epub 2023 Jan 6.
4
Skeleton-Based Action Recognition with Multi-Stream Adaptive Graph Convolutional Networks.基于多流自适应图卷积网络的骨架动作识别
IEEE Trans Image Process. 2020 Oct 9;PP. doi: 10.1109/TIP.2020.3028207.
5
A Comprehensive Survey on Graph Neural Networks.图神经网络综述。
IEEE Trans Neural Netw Learn Syst. 2021 Jan;32(1):4-24. doi: 10.1109/TNNLS.2020.2978386. Epub 2021 Jan 4.
6
EleAtt-RNN: Adding Attentiveness to Neurons in Recurrent Neural Networks.EleAtt-RNN:在循环神经网络中为神经元增添注意力机制
IEEE Trans Image Process. 2019 Sep 2. doi: 10.1109/TIP.2019.2937724.
7
NTU RGB+D 120: A Large-Scale Benchmark for 3D Human Activity Understanding.NTU RGB+D 120:用于三维人体活动理解的大规模基准测试。
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2684-2701. doi: 10.1109/TPAMI.2019.2916873. Epub 2019 May 14.
8
View Adaptive Neural Networks for High Performance Skeleton-Based Human Action Recognition.用于基于骨架的高性能人体动作识别的视图自适应神经网络。
IEEE Trans Pattern Anal Mach Intell. 2019 Aug;41(8):1963-1978. doi: 10.1109/TPAMI.2019.2896631. Epub 2019 Jan 31.
9
Learning Clip Representations for Skeleton-Based 3D Action Recognition.学习基于骨架的 3D 动作识别的剪辑表示。
IEEE Trans Image Process. 2018 Jun;27(6):2842-2855. doi: 10.1109/TIP.2018.2812099.
10
Long short-term memory.长短期记忆
Neural Comput. 1997 Nov 15;9(8):1735-80. doi: 10.1162/neco.1997.9.8.1735.