• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种用于无人机视角目标检测的全局-局部自适应网络。

A Global-Local Self-Adaptive Network for Drone-View Object Detection.

作者信息

Deng Sutao, Li Shuai, Xie Ke, Song Wenfeng, Liao Xiao, Hao Aimin, Qin Hong

出版信息

IEEE Trans Image Process. 2021;30:1556-1569. doi: 10.1109/TIP.2020.3045636. Epub 2021 Jan 5.

DOI:10.1109/TIP.2020.3045636
PMID:33360993
Abstract

Directly benefiting from the deep learning methods, object detection has witnessed a great performance boost in recent years. However, drone-view object detection remains challenging for two main reasons: (1) Objects of tiny-scale with more blurs w.r.t. ground-view objects offer less valuable information towards accurate and robust detection; (2) The unevenly distributed objects make the detection inefficient, especially for regions occupied by crowded objects. Confronting such challenges, we propose an end-to-end global-local self-adaptive network (GLSAN) in this paper. The key components in our GLSAN include a global-local detection network (GLDN), a simple yet efficient self-adaptive region selecting algorithm (SARSA), and a local super-resolution network (LSRN). We integrate a global-local fusion strategy into a progressive scale-varying network to perform more precise detection, where the local fine detector can adaptively refine the target's bounding boxes detected by the global coarse detector via cropping the original images for higher-resolution detection. The SARSA can dynamically crop the crowded regions in the input images, which is unsupervised and can be easily plugged into the networks. Additionally, we train the LSRN to enlarge the cropped images, providing more detailed information for finer-scale feature extraction, helping the detector distinguish foreground and background more easily. The SARSA and LSRN also contribute to data augmentation towards network training, which makes the detector more robust. Extensive experiments and comprehensive evaluations on the VisDrone2019-DET benchmark dataset and UAVDT dataset demonstrate the effectiveness and adaptivity of our method. Towards an industrial application, our network is also applied to a DroneBolts dataset with proven advantages. Our source codes have been available at https://github.com/dengsutao/glsan.

摘要

直接受益于深度学习方法,目标检测近年来取得了显著的性能提升。然而,无人机视角的目标检测仍然具有挑战性,主要有两个原因:(1)与地面视角的目标相比,具有更多模糊的小尺度目标提供的用于精确和稳健检测的有价值信息较少;(2)目标分布不均匀使得检测效率低下,特别是对于被密集目标占据的区域。面对这些挑战,我们在本文中提出了一种端到端的全局-局部自适应网络(GLSAN)。我们的GLSAN中的关键组件包括全局-局部检测网络(GLDN)、一种简单而有效的自适应区域选择算法(SARSA)和一个局部超分辨率网络(LSRN)。我们将全局-局部融合策略集成到一个渐进尺度变化网络中以进行更精确的检测,其中局部精细检测器可以通过裁剪原始图像以进行更高分辨率的检测来自适应地细化全局粗略检测器检测到的目标边界框。SARSA可以动态裁剪输入图像中的密集区域,这是无监督的并且可以很容易地插入到网络中。此外,我们训练LSRN来放大裁剪后的图像,为更精细尺度的特征提取提供更详细的信息,帮助检测器更容易地区分前景和背景。SARSA和LSRN也有助于网络训练的数据增强,这使得检测器更加稳健。在VisDrone2019-DET基准数据集和UAVDT数据集上进行的广泛实验和综合评估证明了我们方法的有效性和适应性。对于工业应用,我们的网络也应用于DroneBolts数据集并显示出优势。我们的源代码已在https://github.com/dengsutao/glsan上提供。

相似文献

1
A Global-Local Self-Adaptive Network for Drone-View Object Detection.一种用于无人机视角目标检测的全局-局部自适应网络。
IEEE Trans Image Process. 2021;30:1556-1569. doi: 10.1109/TIP.2020.3045636. Epub 2021 Jan 5.
2
Towards More Efficient Security Inspection via Deep Learning: A Task-Driven X-ray Image Cropping Scheme.通过深度学习实现更高效的安全检查:一种任务驱动的X射线图像裁剪方案
Micromachines (Basel). 2022 Mar 31;13(4):565. doi: 10.3390/mi13040565.
3
Corner-Point and Foreground-Area IoU Loss: Better Localization of Small Objects in Bounding Box Regression.角点和前景面积交并比损失:在边界框回归中更好地定位小目标。
Sensors (Basel). 2023 May 22;23(10):4961. doi: 10.3390/s23104961.
4
Multi-Scale Feature Integrated Attention-Based Rotation Network for Object Detection in VHR Aerial Images.高分辨率航空图像中目标检测的多尺度特征集成注意力旋转网络。
Sensors (Basel). 2020 Mar 18;20(6):1686. doi: 10.3390/s20061686.
5
Attention CoupleNet: Fully Convolutional Attention Coupling Network for Object Detection.关注 CoupleNet:用于目标检测的全卷积注意力耦合网络。
IEEE Trans Image Process. 2019 Jan;28(1):113-126. doi: 10.1109/TIP.2018.2865280. Epub 2018 Aug 13.
6
FocusDet: an efficient object detector for small object.FocusDet:一种用于小目标的高效目标检测器。
Sci Rep. 2024 May 10;14(1):10697. doi: 10.1038/s41598-024-61136-w.
7
UIU-Net: U-Net in U-Net for Infrared Small Object Detection.UIU-Net:用于红外小目标检测的U-Net嵌套U-Net结构
IEEE Trans Image Process. 2023;32:364-376. doi: 10.1109/TIP.2022.3228497. Epub 2022 Dec 21.
8
Dense Attention Fluid Network for Salient Object Detection in Optical Remote Sensing Images.用于光学遥感图像中显著目标检测的密集注意力流体网络
IEEE Trans Image Process. 2021;30:1305-1317. doi: 10.1109/TIP.2020.3042084. Epub 2020 Dec 23.
9
Collaborative Camouflaged Object Detection: A Large-Scale Dataset and Benchmark.协作式伪装目标检测:一个大规模数据集与基准
IEEE Trans Neural Netw Learn Syst. 2024 Dec;35(12):18470-18484. doi: 10.1109/TNNLS.2023.3317091. Epub 2024 Dec 2.
10
Towards Efficient Detection for Small Objects via Attention-Guided Detection Network and Data Augmentation.通过注意力引导检测网络和数据增强实现小目标的高效检测。
Sensors (Basel). 2022 Oct 9;22(19):7663. doi: 10.3390/s22197663.

引用本文的文献

1
Enhancing UAV Object Detection in Low-Light Conditions with ELS-YOLO: A Lightweight Model Based on Improved YOLOv11.使用ELS-YOLO在低光照条件下增强无人机目标检测:一种基于改进YOLOv11的轻量级模型。
Sensors (Basel). 2025 Jul 17;25(14):4463. doi: 10.3390/s25144463.
2
FocusDet: an efficient object detector for small object.FocusDet:一种用于小目标的高效目标检测器。
Sci Rep. 2024 May 10;14(1):10697. doi: 10.1038/s41598-024-61136-w.
3
FEA-Swin: Foreground Enhancement Attention Swin Transformer Network for Accurate UAV-Based Dense Object Detection.
FEA-Swin:用于基于无人机的精确密集目标检测的前景增强注意力Swin变压器网络
Sensors (Basel). 2022 Sep 15;22(18):6993. doi: 10.3390/s22186993.
4
Lightweight Detection Network Based on Sub-Pixel Convolution and Objectness-Aware Structure for UAV Images.基于子像素卷积和目标感知结构的无人机图像轻量级检测网络
Sensors (Basel). 2021 Aug 22;21(16):5656. doi: 10.3390/s21165656.