• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

Paying more attention on backgrounds: Background-centric attention for UAV detection.

作者信息

Lin Xiuxiu, Niu Yusu, Yu Xinran, Fan Zhun, Zhuang Jiafan, Zou An-Min

机构信息

College of Engineering, Shantou University, Shantou, 515063, China.

University of Electronic Science and Technology of China, Chengdu, 611731, China.

出版信息

Neural Netw. 2025 May;185:107182. doi: 10.1016/j.neunet.2025.107182. Epub 2025 Jan 21.

DOI:10.1016/j.neunet.2025.107182
PMID:39884176
Abstract

Under the advancement of artificial intelligence, Unmanned Aerial Vehicles (UAVs) exhibit efficient flexibility in military reconnaissance, traffic monitoring, and crop analysis. However, the UAV detection faces unique challenges due to the UAV's small size in images, high flight speeds, and limited computational resources. This paper introduces a novel Background-centric Attention Module (BAM) to address these challenges. Unlike traditional methods relying on UAV visual features, the BAM utilizes complex background information to identify UAV presence. The BAM seamlessly integrates into existing UAV detection frameworks, improving accuracy with no significant increase in the computation time. Extensive experiments on challenging datasets, Naval Postgraduate School Drones (NPS), and Flying drones (FLDrones) using mainstream detectors YOLOv5 and TphPlus demonstrate the effectiveness of the BAM in significantly enhancing detection accuracy. This research emphasizes the importance of background information in the UAV detection and proposes a method aligning with human perceptual processes, paving the way for further advancements in the field.

摘要

相似文献

1
Paying more attention on backgrounds: Background-centric attention for UAV detection.
Neural Netw. 2025 May;185:107182. doi: 10.1016/j.neunet.2025.107182. Epub 2025 Jan 21.
2
ASG-YOLOv5: Improved YOLOv5 unmanned aerial vehicle remote sensing aerial images scenario for small object detection based on attention and spatial gating.ASG-YOLOv5:基于注意力和空间门控的改进型 YOLOv5 无人机遥感航空图像场景的小目标检测
PLoS One. 2024 Jun 3;19(6):e0298698. doi: 10.1371/journal.pone.0298698. eCollection 2024.
3
Analysis on security-related concerns of unmanned aerial vehicle: attacks, limitations, and recommendations.分析与无人机安全相关的关注点:攻击、限制因素和建议。
Math Biosci Eng. 2022 Jan 10;19(3):2641-2670. doi: 10.3934/mbe.2022121.
4
A Novel Network Framework on Simultaneous Road Segmentation and Vehicle Detection for UAV Aerial Traffic Images.一种用于无人机空中交通图像的同时进行道路分割和车辆检测的新型网络框架。
Sensors (Basel). 2024 Jun 3;24(11):3606. doi: 10.3390/s24113606.
5
Comparing YOLOv3, YOLOv4 and YOLOv5 for Autonomous Landing Spot Detection in Faulty UAVs.比较 YOLOv3、YOLOv4 和 YOLOv5 在无人机故障自主着陆点检测中的应用。
Sensors (Basel). 2022 Jan 8;22(2):464. doi: 10.3390/s22020464.
6
Visual image design of the internet of things based on AI intelligence.基于人工智能的物联网视觉图像设计
Heliyon. 2023 Nov 25;9(12):e22845. doi: 10.1016/j.heliyon.2023.e22845. eCollection 2023 Dec.
7
Simulating unmanned aerial vehicle flight control and collision detection.模拟无人机飞行控制与碰撞检测。
Vis Comput Ind Biomed Art. 2019 Jun 21;2(1):5. doi: 10.1186/s42492-019-0014-9.
8
A reliable unmanned aerial vehicle multi-ship tracking method.一种可靠的无人机多船跟踪方法。
PLoS One. 2025 Jan 10;20(1):e0316933. doi: 10.1371/journal.pone.0316933. eCollection 2025.
9
Cooperative Location Method for Leader-Follower UAV Formation Based on Follower UAV's Moving Vector.基于跟随无人机运动矢量的领导者-跟随者无人机编队协同定位方法
Sensors (Basel). 2022 Sep 20;22(19):7125. doi: 10.3390/s22197125.
10
Dynamic Object Tracking on Autonomous UAV System for Surveillance Applications.自主无人机系统上的动态目标跟踪用于监控应用。
Sensors (Basel). 2021 Nov 27;21(23):7888. doi: 10.3390/s21237888.