• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

ConMLP:基于 MLP 的自监督对比学习在骨骼数据分析和动作识别中的应用。

ConMLP: MLP-Based Self-Supervised Contrastive Learning for Skeleton Data Analysis and Action Recognition.

机构信息

School of Computing and Engineering, University of Huddersfield, Huddersfield HD1 3DH, UK.

School of Cyberspace Security, Xi'an University of Posts and Telecommunications, Xi'an 710061, China.

出版信息

Sensors (Basel). 2023 Feb 22;23(5):2452. doi: 10.3390/s23052452.

DOI:10.3390/s23052452
PMID:36904656
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10007586/
Abstract

Human action recognition has drawn significant attention because of its importance in computer vision-based applications. Action recognition based on skeleton sequences has rapidly advanced in the last decade. Conventional deep learning-based approaches are based on extracting skeleton sequences through convolutional operations. Most of these architectures are implemented by learning spatial and temporal features through multiple streams. These studies have enlightened the action recognition endeavor from various algorithmic angles. However, three common issues are observed: (1) The models are usually complicated; therefore, they have a correspondingly higher computational complexity. (2) For supervised learning models, the reliance on labels during training is always a drawback. (3) Implementing large models is not beneficial to real-time applications. To address the above issues, in this paper, we propose a multi-layer perceptron (MLP)-based self-supervised learning framework with a contrastive learning loss function (ConMLP). ConMLP does not require a massive computational setup; it can effectively reduce the consumption of computational resources. Compared with supervised learning frameworks, ConMLP is friendly to the huge amount of unlabeled training data. In addition, it has low requirements for system configuration and is more conducive to being embedded in real-world applications. Extensive experiments show that ConMLP achieves the top one inference result of 96.9% on the NTU RGB+D dataset. This accuracy is higher than the state-of-the-art self-supervised learning method. Meanwhile, ConMLP is also evaluated in a supervised learning manner, which has achieved comparable performance to the state of the art of recognition accuracy.

摘要

人体动作识别因其在基于计算机视觉的应用中的重要性而受到广泛关注。基于骨骼序列的动作识别在过去十年中得到了快速发展。传统的基于深度学习的方法基于通过卷积操作提取骨骼序列。这些体系结构中的大多数都是通过通过多个流学习空间和时间特征来实现的。这些研究从各种算法角度启发了动作识别的努力。然而,观察到三个常见问题:(1)模型通常很复杂;因此,它们具有相应更高的计算复杂性。(2)对于监督学习模型,训练期间对标签的依赖始终是一个缺点。(3)实施大型模型不利于实时应用。为了解决上述问题,在本文中,我们提出了一种基于多层感知机(MLP)的自监督学习框架,具有对比学习损失函数(ConMLP)。ConMLP 不需要大量的计算设置;它可以有效地减少计算资源的消耗。与监督学习框架相比,ConMLP 对大量未标记的训练数据更友好。此外,它对系统配置的要求较低,更有利于嵌入到实际应用中。大量实验表明,ConMLP 在 NTU RGB+D 数据集上的推断结果达到了 96.9%的最高水平。这一精度高于最先进的自监督学习方法。同时,ConMLP 也以监督学习的方式进行了评估,其识别精度达到了与最先进方法相当的水平。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b7e3/10007586/8eacc16ebd7a/sensors-23-02452-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b7e3/10007586/8eacc16ebd7a/sensors-23-02452-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b7e3/10007586/8eacc16ebd7a/sensors-23-02452-g001.jpg

相似文献

1
ConMLP: MLP-Based Self-Supervised Contrastive Learning for Skeleton Data Analysis and Action Recognition.ConMLP:基于 MLP 的自监督对比学习在骨骼数据分析和动作识别中的应用。
Sensors (Basel). 2023 Feb 22;23(5):2452. doi: 10.3390/s23052452.
2
Self-Supervised 3D Action Representation Learning With Skeleton Cloud Colorization.基于骨架云彩色化的自监督3D动作表征学习
IEEE Trans Pattern Anal Mach Intell. 2024 Jan;46(1):509-524. doi: 10.1109/TPAMI.2023.3325463. Epub 2023 Dec 5.
3
Local contrastive loss with pseudo-label based self-training for semi-supervised medical image segmentation.基于伪标签自训练的局部对比损失的半监督医学图像分割。
Med Image Anal. 2023 Jul;87:102792. doi: 10.1016/j.media.2023.102792. Epub 2023 Mar 11.
4
DMMG: Dual Min-Max Games for Self-Supervised Skeleton-Based Action Recognition.DMMG:用于基于骨架的自监督动作识别的双最小-最大博弈
IEEE Trans Image Process. 2024;33:395-407. doi: 10.1109/TIP.2023.3338410. Epub 2023 Dec 27.
5
X-Invariant Contrastive Augmentation and Representation Learning for Semi-Supervised Skeleton-Based Action Recognition.用于基于骨架的半监督动作识别的X不变对比增强与表示学习
IEEE Trans Image Process. 2022;31:3852-3867. doi: 10.1109/TIP.2022.3175605. Epub 2022 Jun 2.
6
Multi-Granularity Anchor-Contrastive Representation Learning for Semi-Supervised Skeleton-Based Action Recognition.多粒度锚点对比学习在半监督骨架动作识别中的应用
IEEE Trans Pattern Anal Mach Intell. 2023 Jun;45(6):7559-7576. doi: 10.1109/TPAMI.2022.3222871. Epub 2023 May 5.
7
Contrast-Reconstruction Representation Learning for Self-Supervised Skeleton-Based Action Recognition.用于基于自监督骨架的动作识别的对比重建表示学习
IEEE Trans Image Process. 2022;31:6224-6238. doi: 10.1109/TIP.2022.3207577. Epub 2022 Sep 28.
8
Contrastive self-supervised representation learning without negative samples for multimodal human action recognition.用于多模态人类动作识别的无负样本对比自监督表征学习
Front Neurosci. 2023 Jul 5;17:1225312. doi: 10.3389/fnins.2023.1225312. eCollection 2023.
9
Neighbor-Guided Consistent and Contrastive Learning for Semi-Supervised Action Recognition.基于邻居引导的一致性和对比学习的半监督动作识别。
IEEE Trans Image Process. 2023;32:2215-2227. doi: 10.1109/TIP.2023.3265261. Epub 2023 Apr 18.
10
Cross-view motion consistent self-supervised video inter-intra contrastive for action representation understanding.跨视图运动一致的自我监督视频内-外对比动作表示理解。
Neural Netw. 2024 Nov;179:106578. doi: 10.1016/j.neunet.2024.106578. Epub 2024 Jul 26.

引用本文的文献

1
Recent Advances in Sensor Technology for Healthcare and Biomedical Applications (Volume II).医疗保健和生物医学应用传感器技术的最新进展(第二卷)。
Sensors (Basel). 2023 Jun 27;23(13):5949. doi: 10.3390/s23135949.

本文引用的文献

1
CycleMLP: A MLP-Like Architecture for Dense Visual Predictions.循环多层感知器:用于密集视觉预测的类似多层感知器的架构
IEEE Trans Pattern Anal Mach Intell. 2023 Dec;45(12):14284-14300. doi: 10.1109/TPAMI.2023.3303397. Epub 2023 Nov 3.
2
Contrast-Reconstruction Representation Learning for Self-Supervised Skeleton-Based Action Recognition.用于基于自监督骨架的动作识别的对比重建表示学习
IEEE Trans Image Process. 2022;31:6224-6238. doi: 10.1109/TIP.2022.3207577. Epub 2022 Sep 28.
3
ResMLP: Feedforward Networks for Image Classification With Data-Efficient Training.
ResMLP:具有高效数据训练的图像分类前馈网络。
IEEE Trans Pattern Anal Mach Intell. 2023 Apr;45(4):5314-5321. doi: 10.1109/TPAMI.2022.3206148. Epub 2023 Mar 7.
4
Self-supervised learning in medicine and healthcare.医学和医疗保健中的自我监督学习。
Nat Biomed Eng. 2022 Dec;6(12):1346-1352. doi: 10.1038/s41551-022-00914-1. Epub 2022 Aug 11.
5
On the Post Hoc Explainability of Optimized Self-Organizing Reservoir Network for Action Recognition.优化自组织储层网络的事后可解释性在动作识别中的应用。
Sensors (Basel). 2022 Mar 1;22(5):1905. doi: 10.3390/s22051905.
6
Combining Supervised and Unsupervised Learning Algorithms for Human Activity Recognition.结合监督和无监督学习算法进行人体活动识别。
Sensors (Basel). 2021 Sep 21;21(18):6309. doi: 10.3390/s21186309.
7
A Hierarchical Learning Approach for Human Action Recognition.分层学习方法在人体动作识别中的应用。
Sensors (Basel). 2020 Sep 1;20(17):4946. doi: 10.3390/s20174946.
8
NTU RGB+D 120: A Large-Scale Benchmark for 3D Human Activity Understanding.NTU RGB+D 120:用于三维人体活动理解的大规模基准测试。
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2684-2701. doi: 10.1109/TPAMI.2019.2916873. Epub 2019 May 14.
9
Efficient human pose estimation from single depth images.基于单目深度图像的高效人体姿态估计。
IEEE Trans Pattern Anal Mach Intell. 2013 Dec;35(12):2821-40. doi: 10.1109/TPAMI.2012.241.