• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于视觉的无人机视频中的 HAR 研究:使用直方图和深度学习技术。

Vision-Based HAR in UAV Videos Using Histograms and Deep Learning Techniques.

机构信息

School of Computer Science and Engineering, VIT-AP University, Amaravati 522237, India.

出版信息

Sensors (Basel). 2023 Feb 25;23(5):2569. doi: 10.3390/s23052569.

DOI:10.3390/s23052569
PMID:36904773
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10007408/
Abstract

Activity recognition in unmanned aerial vehicle (UAV) surveillance is addressed in various computer vision applications such as image retrieval, pose estimation, object detection, object detection in videos, object detection in still images, object detection in video frames, face recognition, and video action recognition. In the UAV-based surveillance technology, video segments captured from aerial vehicles make it challenging to recognize and distinguish human behavior. In this research, to recognize a single and multi-human activity using aerial data, a hybrid model of histogram of oriented gradient (HOG), mask-regional convolutional neural network (Mask-RCNN), and bidirectional long short-term memory (Bi-LSTM) is employed. The HOG algorithm extracts patterns, Mask-RCNN extracts feature maps from the raw aerial image data, and the Bi-LSTM network exploits the temporal relationship between the frames for the underlying action in the scene. This Bi-LSTM network reduces the error rate to the greatest extent due to its bidirectional process. This novel architecture generates enhanced segmentation by utilizing the histogram gradient-based instance segmentation and improves the accuracy of classifying human activities using the Bi-LSTM approach. Experimental outcomes demonstrate that the proposed model outperforms the other state-of-the-art models and has achieved 99.25% accuracy on the YouTube-Aerial dataset.

摘要

在各种计算机视觉应用中,如图像检索、姿态估计、目标检测、视频中的目标检测、静态图像中的目标检测、视频帧中的目标检测、人脸识别和视频动作识别,都涉及到无人机 (UAV) 监控中的活动识别。在基于无人机的监控技术中,从空中车辆捕获的视频片段使得识别和区分人类行为变得具有挑战性。在这项研究中,为了使用空中数据识别单人或多人活动,采用了方向梯度直方图 (HOG)、掩模区域卷积神经网络 (Mask-RCNN) 和双向长短时记忆 (Bi-LSTM) 的混合模型。HOG 算法提取模式,Mask-RCNN 从原始空中图像数据中提取特征图,Bi-LSTM 网络利用场景中底层动作的帧之间的时间关系。由于其双向过程,该 Bi-LSTM 网络最大限度地降低了错误率。这种新颖的架构通过利用基于直方图梯度的实例分割来增强分割,并通过 Bi-LSTM 方法提高分类人类活动的准确性。实验结果表明,所提出的模型优于其他最先进的模型,在 YouTube-Aerial 数据集上达到了 99.25%的准确率。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/550a2ceb6371/sensors-23-02569-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/e6c859b5e028/sensors-23-02569-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/a4852f6a4774/sensors-23-02569-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/0cb1705127c5/sensors-23-02569-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/42735ca4335e/sensors-23-02569-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/c4cfdd9958d8/sensors-23-02569-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/c0d669a56eb0/sensors-23-02569-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/bb3cc257cc65/sensors-23-02569-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/70e424cad476/sensors-23-02569-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/cf969664becd/sensors-23-02569-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/0f782bf58869/sensors-23-02569-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/e0770ca60080/sensors-23-02569-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/fb2bd1b5aaf8/sensors-23-02569-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/390a17ba99d6/sensors-23-02569-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/7c57ba67f913/sensors-23-02569-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/4141b8d679db/sensors-23-02569-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/b5ab398c0f13/sensors-23-02569-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/f09750b7a5f8/sensors-23-02569-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/c824d39cef78/sensors-23-02569-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/1a57637977f9/sensors-23-02569-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/550a2ceb6371/sensors-23-02569-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/e6c859b5e028/sensors-23-02569-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/a4852f6a4774/sensors-23-02569-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/0cb1705127c5/sensors-23-02569-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/42735ca4335e/sensors-23-02569-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/c4cfdd9958d8/sensors-23-02569-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/c0d669a56eb0/sensors-23-02569-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/bb3cc257cc65/sensors-23-02569-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/70e424cad476/sensors-23-02569-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/cf969664becd/sensors-23-02569-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/0f782bf58869/sensors-23-02569-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/e0770ca60080/sensors-23-02569-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/fb2bd1b5aaf8/sensors-23-02569-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/390a17ba99d6/sensors-23-02569-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/7c57ba67f913/sensors-23-02569-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/4141b8d679db/sensors-23-02569-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/b5ab398c0f13/sensors-23-02569-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/f09750b7a5f8/sensors-23-02569-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/c824d39cef78/sensors-23-02569-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/1a57637977f9/sensors-23-02569-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/36b7/10007408/550a2ceb6371/sensors-23-02569-g020.jpg

相似文献

1
Vision-Based HAR in UAV Videos Using Histograms and Deep Learning Techniques.基于视觉的无人机视频中的 HAR 研究:使用直方图和深度学习技术。
Sensors (Basel). 2023 Feb 25;23(5):2569. doi: 10.3390/s23052569.
2
Using Deep Learning and Low-Cost RGB and Thermal Cameras to Detect Pedestrians in Aerial Images Captured by Multirotor UAV.利用深度学习以及低成本的 RGB 和热成像摄像机,检测多旋翼无人机航拍图像中的行人。
Sensors (Basel). 2018 Jul 12;18(7):2244. doi: 10.3390/s18072244.
3
Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences.基于方向梯度直方图的动作视频序列中人体动作识别特征融合直方图
Sensors (Basel). 2020 Dec 18;20(24):7299. doi: 10.3390/s20247299.
4
An Efficient Anomaly Recognition Framework Using an Attention Residual LSTM in Surveillance Videos.基于注意力残差 LSTM 的监控视频高效异常识别框架
Sensors (Basel). 2021 Apr 16;21(8):2811. doi: 10.3390/s21082811.
5
Multilevel Structural Components Detection and Segmentation toward Computer Vision-Based Bridge Inspection.面向基于计算机视觉的桥梁检测的多层次结构组件检测与分割。
Sensors (Basel). 2022 May 4;22(9):3502. doi: 10.3390/s22093502.
6
A Novel Network Framework on Simultaneous Road Segmentation and Vehicle Detection for UAV Aerial Traffic Images.一种用于无人机空中交通图像的同时进行道路分割和车辆检测的新型网络框架。
Sensors (Basel). 2024 Jun 3;24(11):3606. doi: 10.3390/s24113606.
7
[Intelligent identification of livestock, a source of infection, based on deep learning of unmanned aerial vehicle images].基于无人机图像深度学习的牲畜感染源智能识别
Zhongguo Xue Xi Chong Bing Fang Zhi Za Zhi. 2023 May 10;35(2):121-127. doi: 10.16250/j.32.1374.2022273.
8
Procapra Przewalskii Tracking Autonomous Unmanned Aerial Vehicle Based on Improved Long and Short-Term Memory Kalman Filters.普氏原羚跟踪自主无人机基于改进的长短期记忆卡尔曼滤波器。
Sensors (Basel). 2023 Apr 13;23(8):3948. doi: 10.3390/s23083948.
9
HIT-UAV: A high-altitude infrared thermal dataset for Unmanned Aerial Vehicle-based object detection.HIT-UAV:基于无人机的目标检测用高空红外热数据集。
Sci Data. 2023 Apr 20;10(1):227. doi: 10.1038/s41597-023-02066-6.
10
CNN-Bi-LSTM: A Complex Environment-Oriented Cattle Behavior Classification Network Based on the Fusion of CNN and Bi-LSTM.CNN-Bi-LSTM:一种基于 CNN 和 Bi-LSTM 融合的复杂环境导向牛行为分类网络。
Sensors (Basel). 2023 Sep 6;23(18):7714. doi: 10.3390/s23187714.

引用本文的文献

1
Unmanned aerial vehicle based multi-person detection via deep neural network models.基于深度神经网络模型的无人机多人检测
Front Neurorobot. 2025 Apr 17;19:1582995. doi: 10.3389/fnbot.2025.1582995. eCollection 2025.

本文引用的文献

1
HIT HAR: Human Image Threshing Machine for Human Activity Recognition Using Deep Learning Models.利用深度学习模型的人体图像过筛机(HIT HAR):用于人体活动识别。
Comput Intell Neurosci. 2022 Oct 6;2022:1808990. doi: 10.1155/2022/1808990. eCollection 2022.
2
Hier R-CNN: Instance-Level Human Parts Detection and A New Benchmark.Hier R-CNN:实例级人体部位检测与新基准。
IEEE Trans Image Process. 2021;30:39-54. doi: 10.1109/TIP.2020.3029901. Epub 2020 Nov 18.
3
Res2Net: A New Multi-Scale Backbone Architecture.Res2Net:一种新的多尺度骨干网络架构。
IEEE Trans Pattern Anal Mach Intell. 2021 Feb;43(2):652-662. doi: 10.1109/TPAMI.2019.2938758. Epub 2021 Jan 8.
4
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
5
Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition.空间金字塔池化在深度卷积网络中的视觉识别。
IEEE Trans Pattern Anal Mach Intell. 2015 Sep;37(9):1904-16. doi: 10.1109/TPAMI.2015.2389824.
6
Fall detection in homes of older adults using the Microsoft Kinect.使用微软Kinect在老年人家庭中进行跌倒检测。
IEEE J Biomed Health Inform. 2015 Jan;19(1):290-301. doi: 10.1109/JBHI.2014.2312180. Epub 2014 Mar 17.
7
Enhanced computer vision with Microsoft Kinect sensor: a review.增强计算机视觉的微软 Kinect 传感器:综述。
IEEE Trans Cybern. 2013 Oct;43(5):1318-34. doi: 10.1109/TCYB.2013.2265378. Epub 2013 Jun 25.
8
Object detection with discriminatively trained part-based models.基于判别式训练的部件模型的目标检测。
IEEE Trans Pattern Anal Mach Intell. 2010 Sep;32(9):1627-45. doi: 10.1109/TPAMI.2009.167.