Suppr超能文献

PIFR:一种用于分析基于姿势角度的人类活动以实现视频中跌倒检测自动化的新方法。

PIFR: A novel approach for analyzing pose angle-based human activity to automate fall detection in videos.

作者信息

Kong Vungsovanreach, Soeng Saravit, Thon Munirot, Cho Wan-Sup, Nayyar Anand, Kim Tae-Kyung

机构信息

Big Data Department, Chungbuk National University, Cheongju-si, Chungcheongbuk-do, South Korea.

Department of Management Information Systems, Chungbuk National University, Cheongju-si, Chungcheongbuk-do, South Korea.

出版信息

PLoS One. 2025 Jun 17;20(6):e0325253. doi: 10.1371/journal.pone.0325253. eCollection 2025.

Abstract

Falls pose a significant health risk for elderly populations, necessitating advanced monitoring technologies. This study introduces a novel two-stage fall detection system that combines computer vision and machine learning to accurately identify fall events. The system uses the YOLOv11 object detection model to track individuals and estimate their body pose by identifying 17 key body points across video frames. The proposed approach extracts nine critical geometric features, including the center of mass and various body angles. These features are used to train a support vector machine (SVM) model for binary classification, distinguishing between standing and lying with high precision. The system's temporal validation method analyzes sequential frame changes, ensuring robust fall detection. Experimental results, evaluated on the University of Rzeszow Fall Detection (URFD) dataset and the Multiple Cameras Fall Dataset (MCFD), demonstrate exceptional performance, achieving 88.8% precision, 94.1% recall, an F1-score of 91.4%, and a specificity of 95.6%. The method outperforms existing approaches by effectively capturing complex geometric changes during fall events. The system is applicable to smart homes, wearable devices, and healthcare monitoring platforms, offering a scalable, reliable, and efficient solution to enhance safety and independence for elderly individuals, thereby contributing to advancements in health-monitoring technology.

摘要

跌倒对老年人群构成重大健康风险,因此需要先进的监测技术。本研究介绍了一种新颖的两阶段跌倒检测系统,该系统结合计算机视觉和机器学习来准确识别跌倒事件。该系统使用YOLOv11目标检测模型来跟踪个体,并通过识别视频帧中的17个关键身体部位来估计他们的身体姿势。所提出的方法提取了九个关键几何特征,包括质心和各种身体角度。这些特征用于训练支持向量机(SVM)模型进行二元分类,高精度地区分站立和躺卧状态。该系统的时间验证方法分析连续帧的变化,确保可靠的跌倒检测。在热舒夫大学跌倒检测(URFD)数据集和多摄像头跌倒数据集(MCFD)上评估的实验结果显示了出色的性能,精度达到88.8%,召回率为94.1%,F1分数为91.4%,特异性为95.6%。该方法通过有效捕捉跌倒事件期间的复杂几何变化,优于现有方法。该系统适用于智能家居、可穿戴设备和医疗监测平台,提供了一种可扩展、可靠且高效的解决方案,以提高老年人的安全性和独立性,从而推动健康监测技术的进步。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9891/12173392/a90d54270625/pone.0325253.g001.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验