• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

带正迁移知识的动态手势识别注意力 3D-Ghost 模块。

Attentive 3D-Ghost Module for Dynamic Hand Gesture Recognition with Positive Knowledge Transfer.

机构信息

Beijing Key Laboratory of Multimedia and Intelligent Software Technology, Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China.

Beijing Research Center for Information Technology, Agriculture, Beijing 100097, China.

出版信息

Comput Intell Neurosci. 2021 Nov 18;2021:5044916. doi: 10.1155/2021/5044916. eCollection 2021.

DOI:10.1155/2021/5044916
PMID:34840561
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8616693/
Abstract

Hand gesture recognition is a challenging topic in the field of computer vision. Multimodal hand gesture recognition based on RGB-D is with higher accuracy than that of only RGB or depth. It is not difficult to conclude that the gain originates from the complementary information existing in the two modalities. However, in reality, multimodal data are not always easy to acquire simultaneously, while unimodal RGB or depth hand gesture data are more general. Therefore, one hand gesture system is expected, in which only unimordal RGB or Depth data is supported for testing, while multimodal RGB-D data is available for training so as to attain the complementary information. Fortunately, a kind of method via multimodal training and unimodal testing has been proposed. However, unimodal feature representation and cross-modality transfer still need to be further improved. To this end, this paper proposes a new 3D-Ghost and Spatial Attention Inflated 3D ConvNet (3DGSAI) to extract high-quality features for each modality. The baseline of 3DGSAI network is Inflated 3D ConvNet (I3D), and two main improvements are proposed. One is 3D-Ghost module, and the other is the spatial attention mechanism. The 3D-Ghost module can extract richer features for hand gesture representation, and the spatial attention mechanism makes the network pay more attention to hand region. This paper also proposes an adaptive parameter for positive knowledge transfer, which ensures that the transfer always occurs from the strong modality network to the weak one. Extensive experiments on SKIG, VIVA, and NVGesture datasets demonstrate that our method is competitive with the state of the art. Especially, the performance of our method reaches 97.87% on the SKIG dataset using only RGB, which is the current best result.

摘要

手势识别是计算机视觉领域的一个具有挑战性的课题。基于 RGB-D 的多模态手势识别比仅基于 RGB 或深度的手势识别具有更高的准确性。不难得出结论,这种增益源于两种模式之间存在的互补信息。然而,在实际中,多模态数据并不总是容易同时获取,而单模态 RGB 或深度手势数据则更为常见。因此,人们期望有一种手势系统,该系统仅支持测试单模态 RGB 或深度数据,而同时可利用多模态 RGB-D 数据进行训练,以获取互补信息。幸运的是,已经提出了一种通过多模态训练和单模态测试的方法。然而,单模态特征表示和跨模态传输仍需要进一步改进。为此,本文提出了一种新的 3D-Ghost 和空间注意力膨胀 3D ConvNet(3DGSAI),以提取每种模式的高质量特征。3DGSAI 网络的基线是膨胀 3D ConvNet(I3D),并提出了两个主要改进。一个是 3D-Ghost 模块,另一个是空间注意力机制。3D-Ghost 模块可以对手势表示提取更丰富的特征,而空间注意力机制使网络对手部区域更加关注。本文还提出了一种用于正知识迁移的自适应参数,该参数确保迁移始终从强模态网络到弱模态网络发生。在 SKIG、VIVA 和 NVGesture 数据集上进行的广泛实验表明,我们的方法具有竞争力。特别是,我们的方法在仅使用 RGB 的 SKIG 数据集上的性能达到 97.87%,这是目前的最佳结果。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/357f152ed3ab/CIN2021-5044916.alg.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/cd0df18a54f2/CIN2021-5044916.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/745c98976fae/CIN2021-5044916.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/7c0e506d30fd/CIN2021-5044916.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/be57a49e6496/CIN2021-5044916.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/9f3892dd3226/CIN2021-5044916.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/e4affbb9a4b3/CIN2021-5044916.006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/357f152ed3ab/CIN2021-5044916.alg.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/cd0df18a54f2/CIN2021-5044916.001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/745c98976fae/CIN2021-5044916.002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/7c0e506d30fd/CIN2021-5044916.003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/be57a49e6496/CIN2021-5044916.004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/9f3892dd3226/CIN2021-5044916.005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/e4affbb9a4b3/CIN2021-5044916.006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e4a7/8616693/357f152ed3ab/CIN2021-5044916.alg.001.jpg

相似文献

1
Attentive 3D-Ghost Module for Dynamic Hand Gesture Recognition with Positive Knowledge Transfer.带正迁移知识的动态手势识别注意力 3D-Ghost 模块。
Comput Intell Neurosci. 2021 Nov 18;2021:5044916. doi: 10.1155/2021/5044916. eCollection 2021.
2
Multi-Scale Attention 3D Convolutional Network for Multimodal Gesture Recognition.用于多模态手势识别的多尺度注意力3D卷积网络
Sensors (Basel). 2022 Mar 21;22(6):2405. doi: 10.3390/s22062405.
3
Dynamic Gesture Recognition Algorithm Based on 3D Convolutional Neural Network.基于三维卷积神经网络的动态手势识别算法。
Comput Intell Neurosci. 2021 Aug 16;2021:4828102. doi: 10.1155/2021/4828102. eCollection 2021.
4
Spatio-temporal continuous gesture recognition under degraded environments: performance comparison between 3D integral imaging (InIm) and RGB-D sensors.退化环境下的时空连续手势识别:3D积分成像(InIm)与RGB-D传感器的性能比较
Opt Express. 2021 Sep 13;29(19):30937-30951. doi: 10.1364/OE.438110.
5
ST-TGR: Spatio-Temporal Representation Learning for Skeleton-Based Teaching Gesture Recognition.ST-TGR:基于骨骼的教学手势识别的时空表征学习
Sensors (Basel). 2024 Apr 18;24(8):2589. doi: 10.3390/s24082589.
6
Deep Dynamic Neural Networks for Multimodal Gesture Segmentation and Recognition.深度动态神经网络用于多模态手势分割与识别。
IEEE Trans Pattern Anal Mach Intell. 2016 Aug;38(8):1583-97. doi: 10.1109/TPAMI.2016.2537340. Epub 2016 Mar 2.
7
CNN Deep Learning with Wavelet Image Fusion of CCD RGB-IR and Depth-Grayscale Sensor Data for Hand Gesture Intention Recognition.CNN 基于 CCD RGB-IR 与深度灰度传感器数据的子波图像融合的深度学习在手势意图识别中的应用。
Sensors (Basel). 2022 Jan 21;22(3):803. doi: 10.3390/s22030803.
8
Architecture Design and VLSI Implementation of 3D Hand Gesture Recognition System.三维手势识别系统的体系结构设计与 VLSI 实现。
Sensors (Basel). 2021 Oct 10;21(20):6724. doi: 10.3390/s21206724.
9
On the Effect of Training Convolution Neural Network for Millimeter-Wave Radar-Based Hand Gesture Recognition.基于毫米波雷达的手势识别中卷积神经网络训练的效果。
Sensors (Basel). 2021 Jan 2;21(1):259. doi: 10.3390/s21010259.
10
Multimode Gesture Recognition Algorithm Based on Convolutional Long Short-Term Memory Network.基于卷积长短期记忆网络的多模式手势识别算法。
Comput Intell Neurosci. 2022 Mar 2;2022:4068414. doi: 10.1155/2022/4068414. eCollection 2022.

本文引用的文献

1
A Fusion Recognition Method Based on Multifeature Hidden Markov Model for Dynamic Hand Gesture.基于多特征隐马尔可夫模型的动态手势融合识别方法。
Comput Intell Neurosci. 2020 Sep 9;2020:8871605. doi: 10.1155/2020/8871605. eCollection 2020.