• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于自拍摄视频的人体手部动作检测与分类的 YOLO 系列。

YOLO Series for Human Hand Action Detection and Classification from Egocentric Videos.

机构信息

Faculty of Engineering Technology, Hung Vuong University, Viet Tri City 35100, Vietnam.

Department of Intelligent Computer Systems, Czestochowa University of Technology, 42-218 Czestochowa, Poland.

出版信息

Sensors (Basel). 2023 Mar 20;23(6):3255. doi: 10.3390/s23063255.

DOI:10.3390/s23063255
PMID:36991971
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10058182/
Abstract

Hand detection and classification is a very important pre-processing step in building applications based on three-dimensional (3D) hand pose estimation and hand activity recognition. To automatically limit the hand data area on egocentric vision (EV) datasets, especially to see the development and performance of the "You Only Live Once" (YOLO) network over the past seven years, we propose a study comparing the efficiency of hand detection and classification based on the YOLO-family networks. This study is based on the following problems: (1) systematizing all architectures, advantages, and disadvantages of YOLO-family networks from version (v)1 to v7; (2) preparing ground-truth data for pre-trained models and evaluation models of hand detection and classification on EV datasets (FPHAB, HOI4D, RehabHand); (3) fine-tuning the hand detection and classification model based on the YOLO-family networks, hand detection, and classification evaluation on the EV datasets. Hand detection and classification results on the YOLOv7 network and its variations were the best across all three datasets. The results of the YOLOv7-w6 network are as follows: FPHAB is = 97% with = 0.5; HOI4D is = 95% with = 0.5; RehabHand is larger than 95% with = 0.5; the processing speed of YOLOv7-w6 is 60 fps with a resolution of 1280 × 1280 pixels and that of YOLOv7 is 133 fps with a resolution of 640 × 640 pixels.

摘要

手检测和分类是基于三维(3D)手姿势估计和手活动识别构建应用程序的非常重要的预处理步骤。为了自动限制基于自我中心视觉(EV)数据集的手数据区域,特别是观察过去七年来“只活一次”(YOLO)网络的发展和性能,我们提出了一项比较基于 YOLO 系列网络的手检测和分类效率的研究。本研究基于以下问题:(1)从版本(v)1 到 v7 系统地整理 YOLO 系列网络的所有架构、优点和缺点;(2)为 EV 数据集(FPHAB、HOI4D、RehabHand)上的预训练模型和手检测与分类评估模型准备地面真实数据;(3)基于 YOLO 系列网络、EV 数据集上手检测和分类评估对手检测和分类模型进行微调。在所有三个数据集上,基于 YOLOv7 网络及其变体的手检测和分类结果均最佳。YOLOv7-w6 网络的结果如下:在 FPHAB 上,=97%,=0.5;在 HOI4D 上,=95%,=0.5;在 RehabHand 上,>95%,=0.5;YOLOv7-w6 的处理速度为 60 fps,分辨率为 1280×1280 像素,而 YOLOv7 的处理速度为 133 fps,分辨率为 640×640 像素。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/ff5d669de560/sensors-23-03255-g024.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/98b937bfca02/sensors-23-03255-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/742b79afa406/sensors-23-03255-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/f67df85c47b2/sensors-23-03255-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/0b739e6ee5e9/sensors-23-03255-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/78a96335b45a/sensors-23-03255-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/e6084b71fa90/sensors-23-03255-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/4495a6c70e2b/sensors-23-03255-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/7c37acc4ab19/sensors-23-03255-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/03504b947c85/sensors-23-03255-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/ec5f522cdf7c/sensors-23-03255-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/7a0d58afe72a/sensors-23-03255-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/5537dc40fae1/sensors-23-03255-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/3dd6ef39f2ad/sensors-23-03255-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/e797b0c1d018/sensors-23-03255-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/23d05087e518/sensors-23-03255-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/93b75c7b3799/sensors-23-03255-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/dad4b3580b2e/sensors-23-03255-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/9f75d2360c68/sensors-23-03255-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/d7eebd24c57c/sensors-23-03255-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/9c682d5dabfc/sensors-23-03255-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/a95b034ae213/sensors-23-03255-g021.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/4a1a2021f466/sensors-23-03255-g022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/88c9054af716/sensors-23-03255-g023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/ff5d669de560/sensors-23-03255-g024.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/98b937bfca02/sensors-23-03255-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/742b79afa406/sensors-23-03255-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/f67df85c47b2/sensors-23-03255-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/0b739e6ee5e9/sensors-23-03255-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/78a96335b45a/sensors-23-03255-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/e6084b71fa90/sensors-23-03255-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/4495a6c70e2b/sensors-23-03255-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/7c37acc4ab19/sensors-23-03255-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/03504b947c85/sensors-23-03255-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/ec5f522cdf7c/sensors-23-03255-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/7a0d58afe72a/sensors-23-03255-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/5537dc40fae1/sensors-23-03255-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/3dd6ef39f2ad/sensors-23-03255-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/e797b0c1d018/sensors-23-03255-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/23d05087e518/sensors-23-03255-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/93b75c7b3799/sensors-23-03255-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/dad4b3580b2e/sensors-23-03255-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/9f75d2360c68/sensors-23-03255-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/d7eebd24c57c/sensors-23-03255-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/9c682d5dabfc/sensors-23-03255-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/a95b034ae213/sensors-23-03255-g021.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/4a1a2021f466/sensors-23-03255-g022.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/88c9054af716/sensors-23-03255-g023.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473a/10058182/ff5d669de560/sensors-23-03255-g024.jpg

相似文献

1
YOLO Series for Human Hand Action Detection and Classification from Egocentric Videos.基于自拍摄视频的人体手部动作检测与分类的 YOLO 系列。
Sensors (Basel). 2023 Mar 20;23(6):3255. doi: 10.3390/s23063255.
2
YOLO Based Breast Masses Detection and Classification in Full-Field Digital Mammograms.基于YOLO的全视野数字乳腺钼靶片中乳腺肿块检测与分类
Comput Methods Programs Biomed. 2021 Mar;200:105823. doi: 10.1016/j.cmpb.2020.105823. Epub 2020 Nov 4.
3
Analysis of Movement and Activities of Handball Players Using Deep Neural Networks.使用深度神经网络对手球运动员的运动和活动进行分析。
J Imaging. 2023 Apr 13;9(4):80. doi: 10.3390/jimaging9040080.
4
EC-YOLO: Improved YOLOv7 Model for PCB Electronic Component Detection.EC-YOLO:用于印刷电路板电子元件检测的改进YOLOv7模型
Sensors (Basel). 2024 Jul 5;24(13):4363. doi: 10.3390/s24134363.
5
Automatic Target Detection from Satellite Imagery Using Machine Learning.基于机器学习的卫星图像自动目标检测。
Sensors (Basel). 2022 Feb 2;22(3):1147. doi: 10.3390/s22031147.
6
Agricultural Greenhouses Detection in High-Resolution Satellite Images Based on Convolutional Neural Networks: Comparison of Faster R-CNN, YOLO v3 and SSD.基于卷积神经网络的高分辨率卫星图像中的农业温室检测:Faster R-CNN、YOLO v3和SSD的比较
Sensors (Basel). 2020 Aug 31;20(17):4938. doi: 10.3390/s20174938.
7
A review on modern defect detection models using DCNNs - Deep convolutional neural networks.基于 DCNN 的现代缺陷检测模型综述 - 深度卷积神经网络。
J Adv Res. 2021 Apr 23;35:33-48. doi: 10.1016/j.jare.2021.03.015. eCollection 2022 Jan.
8
Nature-Inspired Search Method and Custom Waste Object Detection and Classification Model for Smart Waste Bin.受自然启发的搜索方法和定制的废物对象检测与分类模型,用于智能垃圾桶。
Sensors (Basel). 2022 Aug 18;22(16):6176. doi: 10.3390/s22166176.
9
Comparison of RetinaNet, SSD, and YOLO v3 for real-time pill identification.基于 RetinaNet、SSD 和 YOLO v3 的实时药丸识别比较。
BMC Med Inform Decis Mak. 2021 Nov 22;21(1):324. doi: 10.1186/s12911-021-01691-8.
10
YOLO-GW: Quickly and Accurately Detecting Pedestrians in a Foggy Traffic Environment.YOLO-GW:在雾天交通环境中快速准确地检测行人
Sensors (Basel). 2023 Jun 13;23(12):5539. doi: 10.3390/s23125539.

引用本文的文献

1
Deep-learning enabled rapid and low-cost detection of microplastics in consumer products following on-site extraction and image processing.通过现场提取和图像处理,深度学习实现了对消费品中微塑料的快速低成本检测。
RSC Adv. 2025 Apr 4;15(14):10473-10483. doi: 10.1039/d4ra07991d.
2
An improved YOLO v4 used for grape detection in unstructured environment.一种用于非结构化环境中葡萄检测的改进型YOLO v4。
Front Plant Sci. 2023 Jul 13;14:1209910. doi: 10.3389/fpls.2023.1209910. eCollection 2023.

本文引用的文献

1
Analysis of the Hands in Egocentric Vision: A Survey.自我中心视觉中的手分析:调查。
IEEE Trans Pattern Anal Mach Intell. 2023 Jun;45(6):6846-6866. doi: 10.1109/TPAMI.2020.2986648. Epub 2023 May 5.
2
Cascade R-CNN: High Quality Object Detection and Instance Segmentation.级联 R-CNN:高质量目标检测和实例分割。
IEEE Trans Pattern Anal Mach Intell. 2021 May;43(5):1483-1498. doi: 10.1109/TPAMI.2019.2956516. Epub 2021 Apr 1.
3
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
4
Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition.空间金字塔池化在深度卷积网络中的视觉识别。
IEEE Trans Pattern Anal Mach Intell. 2015 Sep;37(9):1904-16. doi: 10.1109/TPAMI.2015.2389824.