• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

无人机注意力:基于无人机摄像头的活动识别中的稀疏加权时间注意力

DroneAttention: Sparse weighted temporal attention for drone-camera based activity recognition.

作者信息

Yadav Santosh Kumar, Luthra Achleshwar, Pahwa Esha, Tiwari Kamlesh, Rathore Heena, Pandey Hari Mohan, Corcoran Peter

机构信息

College of Science and Engineering, National University of Ireland, Galway, H91TK33, Ireland; CogniX, Quadrant-2, 10th Floor, Cyber Towers, Madhapur, Hyderabad, Telangana 500081, India.

Department of CSIS, Birla Institute of Technology and Science Pilani, Pilani Campus, Rajasthan 333031, India.

出版信息

Neural Netw. 2023 Feb;159:57-69. doi: 10.1016/j.neunet.2022.12.005. Epub 2022 Dec 13.

DOI:10.1016/j.neunet.2022.12.005
PMID:36535129
Abstract

Human activity recognition (HAR) using drone-mounted cameras has attracted considerable interest from the computer vision research community in recent years. A robust and efficient HAR system has a pivotal role in fields like video surveillance, crowd behavior analysis, sports analysis, and human-computer interaction. What makes it challenging are the complex poses, understanding different viewpoints, and the environmental scenarios where the action is taking place. To address such complexities, in this paper, we propose a novel Sparse Weighted Temporal Attention (SWTA) module to utilize sparsely sampled video frames for obtaining global weighted temporal attention. The proposed SWTA is comprised of two parts. First, temporal segment network that sparsely samples a given set of frames. Second, weighted temporal attention, which incorporates a fusion of attention maps derived from optical flow, with raw RGB images. This is followed by a basenet network, which comprises a convolutional neural network (CNN) module along with fully connected layers that provide us with activity recognition. The SWTA network can be used as a plug-in module to the existing deep CNN architectures, for optimizing them to learn temporal information by eliminating the need for a separate temporal stream. It has been evaluated on three publicly available benchmark datasets, namely Okutama, MOD20, and Drone-Action. The proposed model has received an accuracy of 72.76%, 92.56%, and 78.86% on the respective datasets thereby surpassing the previous state-of-the-art performances by a margin of 25.26%, 18.56%, and 2.94%, respectively.

摘要

近年来,利用无人机搭载摄像头进行人体活动识别(HAR)引起了计算机视觉研究界的广泛关注。一个强大且高效的HAR系统在视频监控、人群行为分析、体育分析和人机交互等领域起着关键作用。其具有挑战性的地方在于复杂的姿势、对不同视角的理解以及动作发生的环境场景。为了解决这些复杂性问题,在本文中,我们提出了一种新颖的稀疏加权时间注意力(SWTA)模块,以利用稀疏采样的视频帧来获得全局加权时间注意力。所提出的SWTA由两部分组成。第一,时间片段网络,它对给定的一组帧进行稀疏采样。第二,加权时间注意力,它将从光流导出的注意力图与原始RGB图像进行融合。接下来是一个基础网络,它由一个卷积神经网络(CNN)模块以及全连接层组成,为我们提供活动识别。SWTA网络可以用作现有深度CNN架构的插件模块,通过消除对单独时间流的需求来优化它们以学习时间信息。我们在三个公开可用的基准数据集上对其进行了评估,即奥多摩、MOD20和无人机动作数据集。所提出模型在各个数据集上分别取得了72.76%、92.56%和78.86%的准确率,从而分别比之前的最优性能高出25.26%、18.56%和2.94%。

相似文献

1
DroneAttention: Sparse weighted temporal attention for drone-camera based activity recognition.无人机注意力:基于无人机摄像头的活动识别中的稀疏加权时间注意力
Neural Netw. 2023 Feb;159:57-69. doi: 10.1016/j.neunet.2022.12.005. Epub 2022 Dec 13.
2
Two-Level Attention Module Based on Spurious-3D Residual Networks for Human Action Recognition.基于伪 3D 残差网络的两级注意模块的人体动作识别。
Sensors (Basel). 2023 Feb 3;23(3):1707. doi: 10.3390/s23031707.
3
Human Activity Recognition Using Cascaded Dual Attention CNN and Bi-Directional GRU Framework.基于级联双注意力卷积神经网络和双向门控循环单元框架的人类活动识别
J Imaging. 2023 Jun 26;9(7):130. doi: 10.3390/jimaging9070130.
4
A Comprehensive Review of Recent Deep Learning Techniques for Human Activity Recognition.深度学习技术在人体活动识别中的研究进展综述
Comput Intell Neurosci. 2022 Apr 20;2022:8323962. doi: 10.1155/2022/8323962. eCollection 2022.
5
STC-NLSTMNet: An Improved Human Activity Recognition Method Using Convolutional Neural Network with NLSTM from WiFi CSI.STC-NLSTMNet:一种基于 WiFi CSI 的卷积神经网络与 NLSTM 改进的人体活动识别方法。
Sensors (Basel). 2022 Dec 29;23(1):356. doi: 10.3390/s23010356.
6
Vision Transformer and Deep Sequence Learning for Human Activity Recognition in Surveillance Videos.基于视觉Transformer 和深度学习序列模型的监控视频人体行为识别
Comput Intell Neurosci. 2022 Apr 4;2022:3454167. doi: 10.1155/2022/3454167. eCollection 2022.
7
Global and Local Knowledge-Aware Attention Network for Action Recognition.用于动作识别的全局和局部知识感知注意力网络。
IEEE Trans Neural Netw Learn Syst. 2021 Jan;32(1):334-347. doi: 10.1109/TNNLS.2020.2978613. Epub 2021 Jan 4.
8
Visual attention prediction improves performance of autonomous drone racing agents.视觉注意预测能提高自动驾驶无人机竞赛代理的表现。
PLoS One. 2022 Mar 1;17(3):e0264471. doi: 10.1371/journal.pone.0264471. eCollection 2022.
9
Action Recognition by an Attention-Aware Temporal Weighted Convolutional Neural Network.基于注意力感知时间加权卷积神经网络的动作识别。
Sensors (Basel). 2018 Jun 21;18(7):1979. doi: 10.3390/s18071979.
10
Lightweight Semantic-Guided Neural Networks Based on Single Head Attention for Action Recognition.基于单头注意力的轻量级语义引导神经网络在动作识别中的应用。
Sensors (Basel). 2022 Nov 28;22(23):9249. doi: 10.3390/s22239249.

引用本文的文献

1
Deep learning-based action recognition for analyzing drug-induced bone remodeling mechanisms.基于深度学习的动作识别用于分析药物诱导的骨重塑机制。
Front Pharmacol. 2025 May 29;16:1564157. doi: 10.3389/fphar.2025.1564157. eCollection 2025.
2
Unmanned aerial vehicle based multi-person detection via deep neural network models.基于深度神经网络模型的无人机多人检测
Front Neurorobot. 2025 Apr 17;19:1582995. doi: 10.3389/fnbot.2025.1582995. eCollection 2025.
3
Unmanned aerial vehicles for human detection and recognition using neural-network model.
使用神经网络模型进行人体检测与识别的无人机。
Front Neurorobot. 2024 Dec 4;18:1443678. doi: 10.3389/fnbot.2024.1443678. eCollection 2024.
4
Research on the Human Motion Recognition Method Based on Wearable.基于可穿戴设备的人体运动识别方法研究
Biosensors (Basel). 2024 Jul 10;14(7):337. doi: 10.3390/bios14070337.