Suppr超能文献

一种基于计算机视觉的乌尔都语手语数据集识别与分类系统。

A computer vision-based system for recognition and classification of Urdu sign language dataset.

作者信息

Zahid Hira, Rashid Munaf, Syed Sidra Abid, Ullah Rafi, Asif Muhammad, Khan Muzammil, Abdul Mujeeb Amenah, Haider Khan Ali

机构信息

Biomedical Engineering Department and Electrical Engineering Department, Ziauddin University, Karachi, Pakistan.

Electrical Engineering Department and Software Engineering Department, Ziauddin University, Karachi, Pakistan.

出版信息

PeerJ Comput Sci. 2022 Dec 14;8:e1174. doi: 10.7717/peerj-cs.1174. eCollection 2022.

Abstract

Human beings rely heavily on social communication as one of the major aspects of communication. Language is the most effective means of verbal and nonverbal communication and association. To bridge the communication gap between deaf people communities, and non-deaf people, sign language is widely used. According to the World Federation of the Deaf, there are about 70 million deaf people present around the globe and about 300 sign languages being used. Hence, the structural form of the hand gestures involving visual motions and signs is used as a communication system to help the deaf and speech-impaired community for daily interaction. The aim is to collect a dataset of Urdu sign language (USL) and test it through a machine learning classifier. The overview of the proposed system is divided into four main stages , data collection, data acquisition, training model ad testing model. The USL dataset which is comprised of 1,560 images was created by photographing various hand positions using a camera. This work provides a strategy for automated identification of USL numbers based on a bag-of-words (BoW) paradigm. For classification purposes, support vector machine (SVM), Random Forest, and K-nearest neighbor (K-NN) are used with the BoW histogram bin frequencies as characteristics. The proposed technique outperforms others in number classification, attaining the accuracies of 88%, 90%, and 84% for the random forest, SVM, and K-NN respectively.

摘要

人类严重依赖社会交流,将其作为交流的主要方面之一。语言是言语和非言语交流及联系的最有效手段。为了弥合聋人群体与非聋人群体之间的交流差距,手语被广泛使用。据世界聋人联合会统计,全球约有7000万聋人,使用约300种手语。因此,涉及视觉动作和手势的手部姿势的结构形式被用作一种交流系统,以帮助聋人和有语言障碍的群体进行日常互动。目的是收集乌尔都语手语(USL)数据集,并通过机器学习分类器对其进行测试。所提出系统的概述分为四个主要阶段,即数据收集、数据获取、训练模型和测试模型。由1560张图像组成的USL数据集是通过使用相机拍摄各种手部姿势创建的。这项工作提供了一种基于词袋(BoW)范式自动识别USL数字的策略。为了进行分类,将支持向量机(SVM)、随机森林和K近邻(K-NN)与BoW直方图箱频率作为特征一起使用。所提出的技术在数字分类方面优于其他技术,随机森林、SVM和K-NN的准确率分别达到88%、90%和84%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc22/10281630/a92d5c198b22/peerj-cs-08-1174-g001.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验