• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于混合深度学习模型的实时阿拉伯手语识别

Real-Time Arabic Sign Language Recognition Using a Hybrid Deep Learning Model.

机构信息

Department of Computer Science, College of Computer Science and Engineering, Taibah University, Madinah 42353, Saudi Arabia.

出版信息

Sensors (Basel). 2024 Jun 6;24(11):3683. doi: 10.3390/s24113683.

DOI:10.3390/s24113683
PMID:38894473
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11175347/
Abstract

Sign language is an essential means of communication for individuals with hearing disabilities. However, there is a significant shortage of sign language interpreters in some languages, especially in Saudi Arabia. This shortage results in a large proportion of the hearing-impaired population being deprived of services, especially in public places. This paper aims to address this gap in accessibility by leveraging technology to develop systems capable of recognizing Arabic Sign Language (ArSL) using deep learning techniques. In this paper, we propose a hybrid model to capture the spatio-temporal aspects of sign language (i.e., letters and words). The hybrid model consists of a Convolutional Neural Network (CNN) classifier to extract spatial features from sign language data and a Long Short-Term Memory (LSTM) classifier to extract spatial and temporal characteristics to handle sequential data (i.e., hand movements). To demonstrate the feasibility of our proposed hybrid model, we created a dataset of 20 different words, resulting in 4000 images for ArSL: 10 static gesture words and 500 videos for 10 dynamic gesture words. Our proposed hybrid model demonstrates promising performance, with the CNN and LSTM classifiers achieving accuracy rates of 94.40% and 82.70%, respectively. These results indicate that our approach can significantly enhance communication accessibility for the hearing-impaired community in Saudi Arabia. Thus, this paper represents a major step toward promoting inclusivity and improving the quality of life for the hearing impaired.

摘要

手语是听障人士的重要交流方式。然而,某些语言,特别是在沙特阿拉伯,手语翻译人员严重短缺。这导致很大一部分听障人士无法获得服务,尤其是在公共场所。本文旨在通过利用技术开发能够使用深度学习技术识别阿拉伯手语(ArSL)的系统来解决这种无障碍访问的差距。在本文中,我们提出了一种混合模型,用于捕捉手语(即字母和单词)的时空方面。混合模型由卷积神经网络(CNN)分类器组成,用于从手语数据中提取空间特征,以及长短期记忆(LSTM)分类器,用于提取空间和时间特征来处理顺序数据(即手部运动)。为了证明我们提出的混合模型的可行性,我们创建了一个包含 20 个不同单词的数据集,这些单词包括 10 个静态手势词和 500 个动态手势词,共产生了 4000 个 ArSL 图像。我们提出的混合模型表现出了有希望的性能,CNN 和 LSTM 分类器的准确率分别达到了 94.40%和 82.70%。这些结果表明,我们的方法可以极大地增强沙特阿拉伯听障人士的交流无障碍性。因此,本文代表了促进包容性和提高听障人士生活质量的重要一步。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/d815e1c4b18a/sensors-24-03683-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/e66e69c30f29/sensors-24-03683-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/d930ef092af4/sensors-24-03683-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/86491c8de88b/sensors-24-03683-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/13d494286b08/sensors-24-03683-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/9f0a5f81e1b1/sensors-24-03683-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/1aa35a32022e/sensors-24-03683-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/72a28892e87a/sensors-24-03683-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/f2868561d08c/sensors-24-03683-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/d815e1c4b18a/sensors-24-03683-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/e66e69c30f29/sensors-24-03683-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/d930ef092af4/sensors-24-03683-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/86491c8de88b/sensors-24-03683-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/13d494286b08/sensors-24-03683-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/9f0a5f81e1b1/sensors-24-03683-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/1aa35a32022e/sensors-24-03683-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/72a28892e87a/sensors-24-03683-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/f2868561d08c/sensors-24-03683-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/016a/11175347/d815e1c4b18a/sensors-24-03683-g009.jpg

相似文献

1
Real-Time Arabic Sign Language Recognition Using a Hybrid Deep Learning Model.基于混合深度学习模型的实时阿拉伯手语识别
Sensors (Basel). 2024 Jun 6;24(11):3683. doi: 10.3390/s24113683.
2
Efhamni: A Deep Learning-Based Saudi Sign Language Recognition Application.埃法赫尼:一种基于深度学习的沙特手语识别应用。
Sensors (Basel). 2024 May 14;24(10):3112. doi: 10.3390/s24103112.
3
Signer-Independent Arabic Sign Language Recognition System Using Deep Learning Model.基于深度学习模型的无签名者依赖的阿拉伯手语识别系统。
Sensors (Basel). 2023 Aug 14;23(16):7156. doi: 10.3390/s23167156.
4
Sign Language Recognition Using the Electromyographic Signal: A Systematic Literature Review.使用肌电图信号的手语识别:系统文献综述。
Sensors (Basel). 2023 Oct 9;23(19):8343. doi: 10.3390/s23198343.
5
Hypertuned Deep Convolutional Neural Network for Sign Language Recognition.超调深度卷积神经网络的手语识别。
Comput Intell Neurosci. 2022 Apr 30;2022:1450822. doi: 10.1155/2022/1450822. eCollection 2022.
6
Dual Stream Long Short-Term Memory Feature Fusion Classifier for Surface Electromyography Gesture Recognition.双通道长短时记忆特征融合分类器用于表面肌电手势识别。
Sensors (Basel). 2024 Jun 4;24(11):3631. doi: 10.3390/s24113631.
7
American Sign Language Words Recognition of Skeletal Videos Using Processed Video Driven Multi-Stacked Deep LSTM.基于处理视频驱动的多层堆叠深度 LSTM 的骨骼视频的美国手语词识别。
Sensors (Basel). 2022 Feb 11;22(4):1406. doi: 10.3390/s22041406.
8
A comparison of Arabic sign language dynamic gesture recognition models.阿拉伯手语动态手势识别模型的比较
Heliyon. 2020 Mar 14;6(3):e03554. doi: 10.1016/j.heliyon.2020.e03554. eCollection 2020 Mar.
9
Real-Time Hand Gesture Recognition Using Fine-Tuned Convolutional Neural Network.基于微调卷积神经网络的实时手势识别。
Sensors (Basel). 2022 Jan 18;22(3):706. doi: 10.3390/s22030706.
10
Spatial Attention-Based 3D Graph Convolutional Neural Network for Sign Language Recognition.基于空间注意力的 3D 图卷积神经网络的手语识别。
Sensors (Basel). 2022 Jun 16;22(12):4558. doi: 10.3390/s22124558.

引用本文的文献

1
Efficient spatio-temporal modeling for sign language recognition using CNN and RNN architectures.使用卷积神经网络(CNN)和循环神经网络(RNN)架构进行手语识别的高效时空建模。
Front Artif Intell. 2025 Aug 25;8:1630743. doi: 10.3389/frai.2025.1630743. eCollection 2025.
2
Attention-based hybrid deep learning model with CSFOA optimization and G-TverskyUNet3+ for Arabic sign language recognition.基于注意力的混合深度学习模型,采用CSFOA优化和G-TverskyUNet3+进行阿拉伯手语识别。
Sci Rep. 2025 Jun 26;15(1):20313. doi: 10.1038/s41598-025-03560-0.
3
Deep Learning Technology and Image Sensing.

本文引用的文献

1
Sign2Pose: A Pose-Based Approach for Gloss Prediction Using a Transformer Model.Sign2Pose:一种基于姿势的方法,使用转换器模型进行 Gloss 预测。
Sensors (Basel). 2023 Mar 6;23(5):2853. doi: 10.3390/s23052853.
2
Audio-Visual Speech and Gesture Recognition by Sensors of Mobile Devices.基于移动设备传感器的视听语音和手势识别。
Sensors (Basel). 2023 Feb 17;23(4):2284. doi: 10.3390/s23042284.
3
Sign Language Recognition for Arabic Alphabets Using Transfer Learning Technique.基于迁移学习技术的阿拉伯字母手语识别。
深度学习技术与图像传感
Sensors (Basel). 2024 Aug 8;24(16):5130. doi: 10.3390/s24165130.
Comput Intell Neurosci. 2022 Apr 22;2022:4567989. doi: 10.1155/2022/4567989. eCollection 2022.
4
A Review of Recurrent Neural Networks: LSTM Cells and Network Architectures.递归神经网络综述:长短期记忆细胞和网络架构。
Neural Comput. 2019 Jul;31(7):1235-1270. doi: 10.1162/neco_a_01199. Epub 2019 May 21.