• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过轻量级实例分割在边缘设备上检测红外无人机。

Detecting infrared UAVs on edge devices through lightweight instance segmentation.

作者信息

Chen YuZhi, Sun HaoYue, Tian Liang, Yang Ye, Wang ShenYang, Wang TianYou

机构信息

Hebei University of Architecture, Zhangjiakou, China.

Hebei Digital Education Collaborative Innovation, Shijiazhuang, China.

出版信息

PLoS One. 2025 Aug 18;20(8):e0330074. doi: 10.1371/journal.pone.0330074. eCollection 2025.

DOI:10.1371/journal.pone.0330074
PMID:40825008
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12360555/
Abstract

MOTIVATION

Infrared unmanned aerial vehicle (UAV) detection for surveillance applications faces three conflicting requirements: accurate detection of pixel-level thermal signatures, real-time processing capabilities, and deployment feasibility on resource-constrained edge devices. Current deep learning approaches typically optimize for one or two of these objectives while compromising the third.

METHOD

This paper presents YOLO11-AU-IR, a lightweight instance segmentation framework that addresses these challenges through three architectural innovations. First, Efficient Adaptive Downsampling (EADown) employs dual-branch processing with grouped convolutions to preserve small-target spatial features during multi-scale fusion. Second, HeteroScale Attention Network (HSAN) implements grouped multi-scale convolutions with joint channel-spatial attention mechanisms for enhanced cross-scale feature representation. These architectural optimizations collectively reduce computational requirements while maintaining detection accuracy. Third, Adaptive Threshold Focal Loss (ATFL) introduces epoch-adaptive parameter tuning to address the extreme foreground-background imbalance inherent in infrared UAV imagery.

RESULTS

YOLO11-AU-IR is evaluated on the AUVD-Seg300 dataset, achieving 97.7% mAP@0.50 and 75.2% mAP@0.50:0.95, surpassing the YOLO11n-seg baseline by 1.7% and 4.4%, respectively. The model reduces parameters by 24.5% and GFLOPs by 11.8% compared to YOLO11n-seg, while maintaining real-time inference at 59.8 FPS on an NVIDIA RTX 3090 with low variance. On the NVIDIA Jetson TX2, under INT8 CPU-only deployment, YOLO11-AU-IR retains 95% mAP@0.50 with minimal memory footprint and stable performance, demonstrating its practical edge compatibility. Ablation studies further confirm the complementary contributions of EADown, HSAN, and ATFL in enhancing accuracy, robustness, and efficiency. Code and dataset are publicly available at https://github.com/chen-yuzhi/YOLO11-AU-IR.

摘要

动机

用于监视应用的红外无人机(UAV)检测面临三个相互冲突的要求:精确检测像素级热特征、实时处理能力以及在资源受限的边缘设备上的部署可行性。当前的深度学习方法通常针对其中一两个目标进行优化,而牺牲了第三个目标。

方法

本文提出了YOLO11-AU-IR,这是一个轻量级实例分割框架,通过三项架构创新来应对这些挑战。首先,高效自适应下采样(EADown)采用带有分组卷积的双分支处理,以在多尺度融合期间保留小目标空间特征。其次,异尺度注意力网络(HSAN)通过联合通道-空间注意力机制实现分组多尺度卷积,以增强跨尺度特征表示。这些架构优化共同降低了计算需求,同时保持检测精度。第三,自适应阈值焦点损失(ATFL)引入了epoch自适应参数调整,以解决红外无人机图像中固有的极端前景-背景不平衡问题。

结果

YOLO11-AU-IR在AUVD-Seg300数据集上进行评估,在mAP@0.50时达到97.7%,在mAP@0.50:0.95时达到75.2%,分别比YOLO11n-seg基线高出1.7%和4.4%。与YOLO11n-seg相比,该模型的参数减少了24.5%,GFLOP减少了11.8%,同时在NVIDIA RTX 3090上以59.8 FPS保持实时推理,方差较低。在NVIDIA Jetson TX2上,在仅INT8 CPU部署下,YOLO11-AU-IR以最小的内存占用和稳定的性能保持了95%的mAP@0.50,证明了其实际的边缘兼容性。消融研究进一步证实了EADown、HSAN和ATFL在提高准确性、鲁棒性和效率方面的互补作用。代码和数据集可在https://github.com/chen-yuzhi/YOLO11-AU-IR上公开获取。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/c0abf851357b/pone.0330074.g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/334200e2da46/pone.0330074.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/76aa53211aa4/pone.0330074.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/4cc560759ef5/pone.0330074.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/d1b1cab8d9e6/pone.0330074.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/3286c84d4546/pone.0330074.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/51512684306f/pone.0330074.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/f151eb11e001/pone.0330074.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/62eeada6b320/pone.0330074.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/f24dcbce7885/pone.0330074.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/f140551a69b3/pone.0330074.g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/38177e007f92/pone.0330074.g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/eec9b07e137d/pone.0330074.g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/958d18204bde/pone.0330074.g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/c0abf851357b/pone.0330074.g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/334200e2da46/pone.0330074.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/76aa53211aa4/pone.0330074.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/4cc560759ef5/pone.0330074.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/d1b1cab8d9e6/pone.0330074.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/3286c84d4546/pone.0330074.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/51512684306f/pone.0330074.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/f151eb11e001/pone.0330074.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/62eeada6b320/pone.0330074.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/f24dcbce7885/pone.0330074.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/f140551a69b3/pone.0330074.g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/38177e007f92/pone.0330074.g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/eec9b07e137d/pone.0330074.g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/958d18204bde/pone.0330074.g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/537d/12360555/c0abf851357b/pone.0330074.g014.jpg

相似文献

1
Detecting infrared UAVs on edge devices through lightweight instance segmentation.通过轻量级实例分割在边缘设备上检测红外无人机。
PLoS One. 2025 Aug 18;20(8):e0330074. doi: 10.1371/journal.pone.0330074. eCollection 2025.
2
Integrated neural network framework for multi-object detection and recognition using UAV imagery.用于使用无人机图像进行多目标检测与识别的集成神经网络框架。
Front Neurorobot. 2025 Jul 30;19:1643011. doi: 10.3389/fnbot.2025.1643011. eCollection 2025.
3
YOLO-GML: An object edge enhancement detection model for UAV aerial images in complex environments.YOLO-GML:一种用于复杂环境中无人机航空图像的目标边缘增强检测模型。
PLoS One. 2025 Jul 10;20(7):e0328070. doi: 10.1371/journal.pone.0328070. eCollection 2025.
4
DASNet a dual branch multi level attention sheep counting network.DASNet是一种双分支多级注意力羊只计数网络。
Sci Rep. 2025 Jul 2;15(1):23228. doi: 10.1038/s41598-025-97929-w.
5
AC-YOLO: A lightweight ship detection model for SAR images based on YOLO11.AC - YOLO:一种基于YOLO11的用于合成孔径雷达(SAR)图像的轻量级船舶检测模型。
PLoS One. 2025 Jul 30;20(7):e0327362. doi: 10.1371/journal.pone.0327362. eCollection 2025.
6
Multi-level channel-spatial attention and light-weight scale-fusion network (MCSLF-Net): multi-level channel-spatial attention and light-weight scale-fusion transformer for 3D brain tumor segmentation.多级通道空间注意力与轻量级尺度融合网络(MCSLF-Net):用于3D脑肿瘤分割的多级通道空间注意力与轻量级尺度融合变换器
Quant Imaging Med Surg. 2025 Jul 1;15(7):6301-6325. doi: 10.21037/qims-2025-354. Epub 2025 Jun 30.
7
Lightweight cross-resolution coarse-to-fine network for efficient deformable medical image registration.用于高效可变形医学图像配准的轻量级跨分辨率粗到细网络
Med Phys. 2025 Apr 25. doi: 10.1002/mp.17827.
8
Optimized small object detection in low resolution infrared images using super resolution and attention based feature fusion.基于超分辨率和注意力特征融合的低分辨率红外图像中小目标检测优化
PLoS One. 2025 Jul 18;20(7):e0328223. doi: 10.1371/journal.pone.0328223. eCollection 2025.
9
A Pool Drowning Detection Model Based on Improved YOLO.一种基于改进YOLO的泳池溺水检测模型。
Sensors (Basel). 2025 Sep 5;25(17):5552. doi: 10.3390/s25175552.
10
FCMI-YOLO: An efficient deep learning-based algorithm for real-time fire detection on edge devices.FCMI-YOLO:一种基于深度学习的高效算法,用于边缘设备上的实时火灾检测。
PLoS One. 2025 Aug 7;20(8):e0329555. doi: 10.1371/journal.pone.0329555. eCollection 2025.

本文引用的文献

1
MIRA-CAP: Memory-Integrated Retrieval-Augmented Captioning for State-of-the-Art Image and Video Captioning.MIRA-CAP:用于先进图像和视频字幕的内存集成检索增强字幕
Sensors (Basel). 2024 Dec 15;24(24):8013. doi: 10.3390/s24248013.
2
YOLO-G: Improved YOLO for cross-domain object detection.YOLO-G:用于跨域目标检测的改进 YOLO。
PLoS One. 2023 Sep 11;18(9):e0291241. doi: 10.1371/journal.pone.0291241. eCollection 2023.
3
High-Resolution Drone Detection Based on Background Difference and SAG-YOLOv5s.基于背景差分和 SAG-YOLOv5s 的高分辨率无人机检测。
Sensors (Basel). 2022 Aug 4;22(15):5825. doi: 10.3390/s22155825.
4
A dataset for multi-sensor drone detection.一个用于多传感器无人机检测的数据集。
Data Brief. 2021 Oct 27;39:107521. doi: 10.1016/j.dib.2021.107521. eCollection 2021 Dec.
5
Real-Time Small Drones Detection Based on Pruned YOLOv4.基于剪枝 YOLOv4 的实时小型无人机检测。
Sensors (Basel). 2021 May 12;21(10):3374. doi: 10.3390/s21103374.