• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于卷积变换学习的融合框架,用于无人机中尺度不变的长期目标检测与跟踪。

Convolutional transform learning based fusion framework for scale invariant long term target detection and tracking in unmanned aerial vehicles.

作者信息

Alrayes Fatma S, Ahmad Nazir, Alshuhail Asma, Alshammeri Menwa, Alqazzaz Ali, Alkhiri Hassan, Alqurni Jehad Saad, Said Yahia

机构信息

Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, 11671, Riyadh, Saudi Arabia.

Department of Computer Science, Applied College at Mahayil, King Khalid University, Abha, Saudi Arabia.

出版信息

Sci Rep. 2025 Aug 2;15(1):28248. doi: 10.1038/s41598-025-09652-1.

DOI:10.1038/s41598-025-09652-1
PMID:40753260
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12318065/
Abstract

Unmanned aerial vehicles (UAVs) become increasingly available devices with extensive usage as environmental monitoring systems. With the benefit of higher mobility, UAVs are applied to fuel various significant uses in computer vision (CV), providing more effectiveness and accessibility than surveillance cameras with permanent camera view, angle, and scale. Nevertheless, owing to camera motion and composite environments, it is problematic to identify the UAV; conventional models frequently miss UAV detection and make false alarms. Drone-equipped cameras monitor objects at changing altitudes, leading to essential scale variants. The model increases targeted accuracy and decreases false positives using real-time data and machine learning (ML) methods. Its enormous applications range from military operations to urban planning and wildlife monitoring. Therefore, this study develops a novel long-term target detection and tracking model for unmanned aerial vehicles using a deep fusion-based convolutional transform learning (LTTDT-UAVDFCTL) model. The LTTDT-UAVDFCTL model presents a new model to improve the robustness and accuracy of target tracking and detection in scale-variant environments. At first, the presented LTTDT-UAVDFCTL technique performs image pre-processing by utilizing the median median-enhanced wiener filter (MEWF) technique to improve clarity and reduce noise. For object detection (OD), the highly accurate YOLOv8 technique is utilized, followed by feature extraction through a backbone deep fusion-based convolutional transform learning of VGG16, CapsNet, and EfficientNetB7 to capture both spatial and hierarchical features across varying scales. Moreover, the graph convolutional neural network (GCN) technique is employed for long-term target detection and tracking models. Finally, the hybrid nonlinear whale optimization algorithm with sine cosine (SCWOA) is implemented for the optimum choice of the hyperparameters involved in the GCN technique. The experimental study of the LTTDT-UAVDFCTL approach is performed under the VisDrone dataset. The performance validation of the LTTDT-UAVDFCTL approach portrayed a superior mAP value of 80.13% over existing models.

摘要

无人驾驶飞行器(UAV)作为环境监测系统,正成为越来越容易获得且用途广泛的设备。受益于更高的机动性,无人机被应用于计算机视觉(CV)中的各种重要用途,与具有固定视角、角度和比例的监控摄像头相比,具有更高的效率和可及性。然而,由于相机运动和复杂环境,识别无人机存在问题;传统模型经常错过无人机检测并产生误报。配备无人机的相机在不同高度监测物体,导致关键的比例变化。该模型使用实时数据和机器学习(ML)方法提高了目标准确性并减少了误报。其广泛的应用范围涵盖军事行动、城市规划和野生动物监测。因此,本研究开发了一种基于深度融合的卷积变换学习的新型无人机长期目标检测与跟踪模型(LTTDT-UAVDFCTL)。LTTDT-UAVDFCTL模型提出了一种新模型,以提高在比例变化环境中目标跟踪和检测的鲁棒性和准确性。首先,所提出的LTTDT-UAVDFCTL技术通过利用中值中值增强维纳滤波器(MEWF)技术进行图像预处理,以提高清晰度并减少噪声。对于目标检测(OD),采用了高精度的YOLOv8技术,随后通过基于VGG16、CapsNet和EfficientNetB7的骨干深度融合卷积变换学习进行特征提取,以跨不同比例捕获空间和层次特征。此外,图卷积神经网络(GCN)技术用于长期目标检测和跟踪模型。最后,实现了具有正弦余弦的混合非线性鲸鱼优化算法(SCWOA),以优化GCN技术中涉及的超参数选择。LTTDT-UAVDFCTL方法的实验研究是在VisDrone数据集下进行的。LTTDT-UAVDFCTL方法的性能验证表明,其平均精度均值(mAP)值比现有模型高出80.13%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/44cfa36c942c/41598_2025_9652_Fig20_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/8604bab68a6d/41598_2025_9652_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/dedda5c3fa48/41598_2025_9652_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/d398696795e8/41598_2025_9652_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/3ab9d2eb55b6/41598_2025_9652_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/a5540316ffc1/41598_2025_9652_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/a71f0d1592dc/41598_2025_9652_Figb_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/1b477147a182/41598_2025_9652_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/56f6d837680e/41598_2025_9652_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/2b07c5333dde/41598_2025_9652_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/0385d0849d93/41598_2025_9652_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/6afca9c8dbdc/41598_2025_9652_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/3007f62943f9/41598_2025_9652_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/41213511ee9e/41598_2025_9652_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/f99610b6457e/41598_2025_9652_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/93ebe9117d9e/41598_2025_9652_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/783f0fb872e7/41598_2025_9652_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/39d726663bb1/41598_2025_9652_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/f7a50c44de9d/41598_2025_9652_Fig17_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/921e060d5bdb/41598_2025_9652_Fig18_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/86f8ca5b9d7b/41598_2025_9652_Fig19_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/44cfa36c942c/41598_2025_9652_Fig20_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/8604bab68a6d/41598_2025_9652_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/dedda5c3fa48/41598_2025_9652_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/d398696795e8/41598_2025_9652_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/3ab9d2eb55b6/41598_2025_9652_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/a5540316ffc1/41598_2025_9652_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/a71f0d1592dc/41598_2025_9652_Figb_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/1b477147a182/41598_2025_9652_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/56f6d837680e/41598_2025_9652_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/2b07c5333dde/41598_2025_9652_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/0385d0849d93/41598_2025_9652_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/6afca9c8dbdc/41598_2025_9652_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/3007f62943f9/41598_2025_9652_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/41213511ee9e/41598_2025_9652_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/f99610b6457e/41598_2025_9652_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/93ebe9117d9e/41598_2025_9652_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/783f0fb872e7/41598_2025_9652_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/39d726663bb1/41598_2025_9652_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/f7a50c44de9d/41598_2025_9652_Fig17_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/921e060d5bdb/41598_2025_9652_Fig18_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/86f8ca5b9d7b/41598_2025_9652_Fig19_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/1220/12318065/44cfa36c942c/41598_2025_9652_Fig20_HTML.jpg

相似文献

1
Convolutional transform learning based fusion framework for scale invariant long term target detection and tracking in unmanned aerial vehicles.基于卷积变换学习的融合框架,用于无人机中尺度不变的长期目标检测与跟踪。
Sci Rep. 2025 Aug 2;15(1):28248. doi: 10.1038/s41598-025-09652-1.
2
An efficient privacy-preserving multilevel fusion-based feature engineering framework for UAV-enabled land cover classification using remote sensing images.一种基于高效隐私保护多级融合的特征工程框架,用于利用遥感图像进行无人机支持的土地覆盖分类。
Sci Rep. 2025 Jul 3;15(1):23707. doi: 10.1038/s41598-025-08930-2.
3
Multi class aerial image classification in UAV networks employing Snake Optimization Algorithm with Deep Learning.基于深度学习并采用蛇优化算法的无人机网络中的多类航空图像分类
Sci Rep. 2025 Jul 4;15(1):23872. doi: 10.1038/s41598-025-04570-8.
4
Gesture recognition for hearing impaired people using an ensemble of deep learning models with improving beluga whale optimization-based hyperparameter tuning.基于改进的白鲸优化超参数调优的深度学习模型集成用于听力障碍者的手势识别
Sci Rep. 2025 Jul 1;15(1):21441. doi: 10.1038/s41598-025-06680-9.
5
DASNet a dual branch multi level attention sheep counting network.DASNet是一种双分支多级注意力羊只计数网络。
Sci Rep. 2025 Jul 2;15(1):23228. doi: 10.1038/s41598-025-97929-w.
6
Short-Term Memory Impairment短期记忆障碍
7
YOLO-GML: An object edge enhancement detection model for UAV aerial images in complex environments.YOLO-GML:一种用于复杂环境中无人机航空图像的目标边缘增强检测模型。
PLoS One. 2025 Jul 10;20(7):e0328070. doi: 10.1371/journal.pone.0328070. eCollection 2025.
8
Advanced object detection for smart accessibility: a Yolov10 with marine predator algorithm to aid visually challenged people.用于智能无障碍的先进目标检测:一种结合海洋捕食者算法的Yolov10,以帮助视障人士。
Sci Rep. 2025 Jul 1;15(1):20759. doi: 10.1038/s41598-025-04959-5.
9
High-Accuracy Polymer Property Detection via Pareto-Optimized SMILES-Based Deep Learning.通过帕累托优化的基于SMILES的深度学习实现高精度聚合物性能检测。
Polymers (Basel). 2025 Jun 28;17(13):1801. doi: 10.3390/polym17131801.
10
A hybrid model for detecting motion artifacts in ballistocardiogram signals.一种用于检测心冲击图信号中运动伪影的混合模型。
Biomed Eng Online. 2025 Jul 23;24(1):92. doi: 10.1186/s12938-025-01426-0.

本文引用的文献

1
UAV target tracking method based on global feature interaction and anchor-frame-free perceptual feature modulation.基于全局特征交互和无锚框感知特征调制的无人机目标跟踪方法
PLoS One. 2025 Jan 16;20(1):e0314485. doi: 10.1371/journal.pone.0314485. eCollection 2025.
2
A Hybrid Nonlinear Whale Optimization Algorithm with Sine Cosine for Global Optimization.一种用于全局优化的基于正弦余弦的混合非线性鲸鱼优化算法。
Biomimetics (Basel). 2024 Oct 7;9(10):602. doi: 10.3390/biomimetics9100602.
3
Early cancer detection using deep learning and medical imaging: A survey.
基于深度学习和医学成像的早期癌症检测:综述。
Crit Rev Oncol Hematol. 2024 Dec;204:104528. doi: 10.1016/j.critrevonc.2024.104528. Epub 2024 Oct 15.
4
Enhancing automated vehicle identification by integrating YOLO v8 and OCR techniques for high-precision license plate detection and recognition.通过集成YOLO v8和OCR技术来增强自动车辆识别,以实现高精度车牌检测与识别。
Sci Rep. 2024 Jun 22;14(1):14389. doi: 10.1038/s41598-024-65272-1.
5
White shark optimizer with optimal deep learning based effective unmanned aerial vehicles communication and scene classification.基于最优深度学习的白鲨优化器在无人机有效通信与场景分类中的应用
Sci Rep. 2023 Dec 27;13(1):23041. doi: 10.1038/s41598-023-50064-w.
6
UAV-YOLOv8: A Small-Object-Detection Model Based on Improved YOLOv8 for UAV Aerial Photography Scenarios.无人机 - YOLOv8:一种基于改进YOLOv8的用于无人机航拍场景的小目标检测模型。
Sensors (Basel). 2023 Aug 15;23(16):7190. doi: 10.3390/s23167190.
7
Detection and Tracking Meet Drones Challenge.检测与跟踪遭遇无人机挑战。
IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):7380-7399. doi: 10.1109/TPAMI.2021.3119563. Epub 2022 Oct 4.