• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于卷积神经网络的人类动作识别中的多级特征融合:以EfficientNet-B7为例

Multi-Level Feature Fusion in CNN-Based Human Action Recognition: A Case Study on EfficientNet-B7.

作者信息

Lueangwitchajaroen Pitiwat, Watcharapinchai Sitapa, Tepsan Worawit, Sooksatra Sorn

机构信息

National Electronic and Computer Technology Center, National Science and Technology Development Agency, Khlong Luang, Pathum Thani 12120, Thailand.

International College of Digital Innovation, Chiang Mai University, Mueang Chiang Mai, Chiang Mai 50200, Thailand.

出版信息

J Imaging. 2024 Dec 12;10(12):320. doi: 10.3390/jimaging10120320.

DOI:10.3390/jimaging10120320
PMID:39728217
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11677249/
Abstract

Accurate human action recognition is becoming increasingly important across various fields, including healthcare and self-driving cars. A simple approach to enhance model performance is incorporating additional data modalities, such as depth frames, point clouds, and skeleton information, while previous studies have predominantly used late fusion techniques to combine these modalities, our research introduces a multi-level fusion approach that combines information at early, intermediate, and late stages together. Furthermore, recognizing the challenges of collecting multiple data types in real-world applications, our approach seeks to exploit multimodal techniques while relying solely on RGB frames as the single data source. In our work, we used RGB frames from the NTU RGB+D dataset as the sole data source. From these frames, we extracted 2D skeleton coordinates and optical flow frames using pre-trained models. We evaluated our multi-level fusion approach with EfficientNet-B7 as a case study, and our methods demonstrated significant improvement, achieving 91.5% in NTU RGB+D 60 dataset accuracy compared to single-modality and single-view models. Despite their simplicity, our methods are also comparable to other state-of-the-art approaches.

摘要

准确的人类行为识别在包括医疗保健和自动驾驶汽车在内的各个领域正变得越来越重要。一种提高模型性能的简单方法是纳入额外的数据模态,如深度帧、点云及骨骼信息。虽然先前的研究主要使用后期融合技术来组合这些模态,但我们的研究引入了一种多级别融合方法,该方法将早期、中期和后期阶段的信息结合在一起。此外,认识到在实际应用中收集多种数据类型的挑战,我们的方法旨在利用多模态技术,同时仅依赖RGB帧作为单一数据源。在我们的工作中,我们使用来自NTU RGB+D数据集的RGB帧作为唯一数据源。从这些帧中,我们使用预训练模型提取了二维骨骼坐标和光流帧。我们以EfficientNet-B7为例评估了我们的多级别融合方法,我们的方法显示出显著改进,在NTU RGB+D 60数据集中相比单模态和单视图模型准确率达到了91.5%。尽管我们的方法很简单,但也可与其他最先进的方法相媲美。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e06/11677249/16fe3e201af9/jimaging-10-00320-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e06/11677249/06a0b0e119f0/jimaging-10-00320-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e06/11677249/cadebe3b9db8/jimaging-10-00320-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e06/11677249/16fe3e201af9/jimaging-10-00320-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e06/11677249/06a0b0e119f0/jimaging-10-00320-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e06/11677249/cadebe3b9db8/jimaging-10-00320-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e06/11677249/16fe3e201af9/jimaging-10-00320-g003.jpg

相似文献

1
Multi-Level Feature Fusion in CNN-Based Human Action Recognition: A Case Study on EfficientNet-B7.基于卷积神经网络的人类动作识别中的多级特征融合:以EfficientNet-B7为例
J Imaging. 2024 Dec 12;10(12):320. doi: 10.3390/jimaging10120320.
2
MMNet: A Model-Based Multimodal Network for Human Action Recognition in RGB-D Videos.MMNet:一种基于模型的 RGB-D 视频人体动作识别多模态网络。
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3522-3538. doi: 10.1109/TPAMI.2022.3177813. Epub 2023 Feb 3.
3
A discriminative multi-modal adaptation neural network model for video action recognition.一种用于视频动作识别的判别式多模态自适应神经网络模型。
Neural Netw. 2025 May;185:107114. doi: 10.1016/j.neunet.2024.107114. Epub 2025 Jan 3.
4
Multi-Modality Adaptive Feature Fusion Graph Convolutional Network for Skeleton-Based Action Recognition.基于骨架的动作识别的多模态自适应特征融合图卷积网络。
Sensors (Basel). 2023 Jun 7;23(12):5414. doi: 10.3390/s23125414.
5
Two-stream spatio-temporal GCN-transformer networks for skeleton-based action recognition.用于基于骨架的动作识别的双流时空GCN-Transformer网络
Sci Rep. 2025 Feb 10;15(1):4982. doi: 10.1038/s41598-025-87752-8.
6
MAF-Net: A multimodal data fusion approach for human action recognition.MAF-Net:一种用于人类动作识别的多模态数据融合方法。
PLoS One. 2025 Apr 9;20(4):e0319656. doi: 10.1371/journal.pone.0319656. eCollection 2025.
7
Using Direct Acyclic Graphs to Enhance Skeleton-Based Action Recognition with a Linear-Map Convolution Neural Network.基于有向无环图的线性映射卷积神经网络在骨骼动作识别中的应用。
Sensors (Basel). 2021 Apr 29;21(9):3112. doi: 10.3390/s21093112.
8
The effect of depth data and upper limb impairment on lightweight monocular RGB human pose estimation models.深度数据和上肢损伤对轻量级单目RGB人体姿态估计模型的影响。
Biomed Eng Online. 2025 Feb 7;24(1):12. doi: 10.1186/s12938-025-01347-y.
9
Estimating strawberry weight for grading by picking robot with point cloud completion and multimodal fusion network.利用点云补全和多模态融合网络的采摘机器人对草莓重量进行分级估计。
Sci Rep. 2025 Apr 2;15(1):11227. doi: 10.1038/s41598-025-92641-1.
10
A Deep Sequence Learning Framework for Action Recognition in Small-Scale Depth Video Dataset.用于小规模深度视频数据集动作识别的深度序列学习框架。
Sensors (Basel). 2022 Sep 9;22(18):6841. doi: 10.3390/s22186841.

本文引用的文献

1
MMNet: A Model-Based Multimodal Network for Human Action Recognition in RGB-D Videos.MMNet:一种基于模型的 RGB-D 视频人体动作识别多模态网络。
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3522-3538. doi: 10.1109/TPAMI.2022.3177813. Epub 2023 Feb 3.
2
VPN++: Rethinking Video-Pose Embeddings for Understanding Activities of Daily Living.VPN++:重新思考视频姿态嵌入以理解日常生活活动。
IEEE Trans Pattern Anal Mach Intell. 2022 Dec;44(12):9703-9717. doi: 10.1109/TPAMI.2021.3127885. Epub 2022 Nov 7.
3
Image Processing Technique and Hidden Markov Model for an Elderly Care Monitoring System.
用于老年护理监测系统的图像处理技术与隐马尔可夫模型
J Imaging. 2020 Jun 13;6(6):49. doi: 10.3390/jimaging6060049.
4
Multimodal Medical Supervised Image Fusion Method by CNN.基于卷积神经网络的多模态医学监督图像融合方法
Front Neurosci. 2021 Jun 2;15:638976. doi: 10.3389/fnins.2021.638976. eCollection 2021.
5
NTU RGB+D 120: A Large-Scale Benchmark for 3D Human Activity Understanding.NTU RGB+D 120:用于三维人体活动理解的大规模基准测试。
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2684-2701. doi: 10.1109/TPAMI.2019.2916873. Epub 2019 May 14.
6
Long short-term memory.长短期记忆
Neural Comput. 1997 Nov 15;9(8):1735-80. doi: 10.1162/neco.1997.9.8.1735.