• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

使用神经网络模型进行人体检测与识别的无人机。

Unmanned aerial vehicles for human detection and recognition using neural-network model.

作者信息

Abbas Yawar, Al Mudawi Naif, Alabdullah Bayan, Sadiq Touseef, Algarni Asaad, Rahman Hameedur, Jalal Ahmad

机构信息

Faculty of Computer Science and AI, Air University, Islamabad, Pakistan.

Department of Computer Science, College of Computer Science and Information System, Najran University, Najran, Saudi Arabia.

出版信息

Front Neurorobot. 2024 Dec 4;18:1443678. doi: 10.3389/fnbot.2024.1443678. eCollection 2024.

DOI:10.3389/fnbot.2024.1443678
PMID:39698500
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11652500/
Abstract

INTRODUCTION

Recognizing human actions is crucial for allowing machines to understand and recognize human behavior, with applications spanning video based surveillance systems, human-robot collaboration, sports analysis systems, and entertainment. The immense diversity in human movement and appearance poses a significant challenge in this field, especially when dealing with drone-recorded (RGB) videos. Factors such as dynamic backgrounds, motion blur, occlusions, varying video capture angles, and exposure issues greatly complicate recognition tasks.

METHODS

In this study, we suggest a method that addresses these challenges in RGB videos captured by drones. Our approach begins by segmenting the video into individual frames, followed by preprocessing steps applied to these RGB frames. The preprocessing aims to reduce computational costs, optimize image quality, and enhance foreground objects while removing the background.

RESULT

This results in improved visibility of foreground objects while eliminating background noise. Next, we employ the YOLOv9 detection algorithm to identify human bodies within the images. From the grayscale silhouette, we extract the human skeleton and identify 15 important locations, such as the head, neck, shoulders (left and right), elbows, wrists, hips, knees, ankles, and hips (left and right), and belly button. By using all these points, we extract specific positions, angular and distance relationships between them, as well as 3D point clouds and fiducial points. Subsequently, we optimize this data using the kernel discriminant analysis (KDA) optimizer, followed by classification using a deep neural network (CNN). To validate our system, we conducted experiments on three benchmark datasets: UAV-Human, UCF, and Drone-Action.

DISCUSSION

On these datasets, our suggested model produced corresponding action recognition accuracies of 0.68, 0.75, and 0.83.

摘要

引言

识别人类行为对于使机器理解和识别人类行为至关重要,其应用涵盖基于视频的监控系统、人机协作、体育分析系统和娱乐领域。人类运动和外观的巨大多样性给该领域带来了重大挑战,尤其是在处理无人机录制的(RGB)视频时。动态背景、运动模糊、遮挡、不同的视频拍摄角度和曝光问题等因素使识别任务变得极为复杂。

方法

在本研究中,我们提出了一种方法来应对无人机拍摄的RGB视频中的这些挑战。我们的方法首先将视频分割成单个帧,然后对这些RGB帧应用预处理步骤。预处理旨在降低计算成本、优化图像质量、增强前景物体同时去除背景。

结果

这使得前景物体的可见性得到改善,同时消除了背景噪声。接下来,我们使用YOLOv9检测算法在图像中识别人体。从灰度轮廓中,我们提取人体骨骼并识别15个重要位置,如头部、颈部、肩膀(左右)、肘部、手腕、臀部、膝盖、脚踝以及臀部(左右)和肚脐。通过使用所有这些点,我们提取它们之间的特定位置、角度和距离关系,以及3D点云和基准点。随后,我们使用核判别分析(KDA)优化器对这些数据进行优化,然后使用深度神经网络(CNN)进行分类。为了验证我们的系统,我们在三个基准数据集上进行了实验:无人机 - 人类、UCF和无人机 - 动作。

讨论

在这些数据集上,我们提出的模型产生的相应动作识别准确率分别为0.68、0.75和0.83。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/5fa374d72c86/fnbot-18-1443678-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/213636cc5829/fnbot-18-1443678-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/50f93ecfe4ac/fnbot-18-1443678-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/ef1152f739cf/fnbot-18-1443678-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/c66505e6e496/fnbot-18-1443678-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/ceaa30d2576b/fnbot-18-1443678-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/18adddf957ae/fnbot-18-1443678-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/8b041e6f3bc9/fnbot-18-1443678-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/a21a45fbbd28/fnbot-18-1443678-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/3a4fd90f5546/fnbot-18-1443678-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/8a8dffdce12b/fnbot-18-1443678-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/0df8eeeae9a8/fnbot-18-1443678-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/8a95c5c6b646/fnbot-18-1443678-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/2f144df2cfc0/fnbot-18-1443678-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/5fa374d72c86/fnbot-18-1443678-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/213636cc5829/fnbot-18-1443678-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/50f93ecfe4ac/fnbot-18-1443678-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/ef1152f739cf/fnbot-18-1443678-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/c66505e6e496/fnbot-18-1443678-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/ceaa30d2576b/fnbot-18-1443678-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/18adddf957ae/fnbot-18-1443678-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/8b041e6f3bc9/fnbot-18-1443678-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/a21a45fbbd28/fnbot-18-1443678-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/3a4fd90f5546/fnbot-18-1443678-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/8a8dffdce12b/fnbot-18-1443678-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/0df8eeeae9a8/fnbot-18-1443678-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/8a95c5c6b646/fnbot-18-1443678-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/2f144df2cfc0/fnbot-18-1443678-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/cc42/11652500/5fa374d72c86/fnbot-18-1443678-g014.jpg

相似文献

1
Unmanned aerial vehicles for human detection and recognition using neural-network model.使用神经网络模型进行人体检测与识别的无人机。
Front Neurorobot. 2024 Dec 4;18:1443678. doi: 10.3389/fnbot.2024.1443678. eCollection 2024.
2
Using Deep Learning and Low-Cost RGB and Thermal Cameras to Detect Pedestrians in Aerial Images Captured by Multirotor UAV.利用深度学习以及低成本的 RGB 和热成像摄像机,检测多旋翼无人机航拍图像中的行人。
Sensors (Basel). 2018 Jul 12;18(7):2244. doi: 10.3390/s18072244.
3
Target detection and classification via EfficientDet and CNN over unmanned aerial vehicles.通过高效检测器(EfficientDet)和卷积神经网络(CNN)在无人机上进行目标检测与分类
Front Neurorobot. 2024 Aug 30;18:1448538. doi: 10.3389/fnbot.2024.1448538. eCollection 2024.
4
Unmanned aerial vehicle based multi-person detection via deep neural network models.基于深度神经网络模型的无人机多人检测
Front Neurorobot. 2025 Apr 17;19:1582995. doi: 10.3389/fnbot.2025.1582995. eCollection 2025.
5
Real-Time and Accurate Drone Detection in a Video with a Static Background.实时且精确的静态背景视频中无人机检测
Sensors (Basel). 2020 Jul 10;20(14):3856. doi: 10.3390/s20143856.
6
Drone Model Classification Using Convolutional Neural Network Trained on Synthetic Data.基于合成数据训练的卷积神经网络的无人机模型分类
J Imaging. 2022 Aug 12;8(8):218. doi: 10.3390/jimaging8080218.
7
Deep Convolutional Neural Network for Flood Extent Mapping Using Unmanned Aerial Vehicles Data.基于无人机数据的深度卷积神经网络洪水淹没范围制图
Sensors (Basel). 2019 Mar 27;19(7):1486. doi: 10.3390/s19071486.
8
Vision-Based HAR in UAV Videos Using Histograms and Deep Learning Techniques.基于视觉的无人机视频中的 HAR 研究:使用直方图和深度学习技术。
Sensors (Basel). 2023 Feb 25;23(5):2569. doi: 10.3390/s23052569.
9
Face mask identification with enhanced cuckoo optimization and deep learning-based faster regional neural network.基于增强型布谷鸟优化和深度学习的更快区域神经网络的面部口罩识别。
Sci Rep. 2024 Nov 29;14(1):29719. doi: 10.1038/s41598-024-78746-z.
10
A dataset for multi-sensor drone detection.一个用于多传感器无人机检测的数据集。
Data Brief. 2021 Oct 27;39:107521. doi: 10.1016/j.dib.2021.107521. eCollection 2021 Dec.

本文引用的文献

1
DroneAttention: Sparse weighted temporal attention for drone-camera based activity recognition.无人机注意力:基于无人机摄像头的活动识别中的稀疏加权时间注意力
Neural Netw. 2023 Feb;159:57-69. doi: 10.1016/j.neunet.2022.12.005. Epub 2022 Dec 13.
2
MSST-RT: Multi-Stream Spatial-Temporal Relative Transformer for Skeleton-Based Action Recognition.基于骨架的动作识别的多流时空相对Transformer(MSST-RT):Multi-Stream Spatial-Temporal Relative Transformer for Skeleton-Based Action Recognition。
Sensors (Basel). 2021 Aug 7;21(16):5339. doi: 10.3390/s21165339.