• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

YOLOv8-MU:一种基于大内核模块和多分支重参数化模块的改进型YOLOv8水下探测器。

YOLOv8-MU: An Improved YOLOv8 Underwater Detector Based on a Large Kernel Block and a Multi-Branch Reparameterization Module.

作者信息

Jiang Xing, Zhuang Xiting, Chen Jisheng, Zhang Jian, Zhang Yiwen

机构信息

School of Tropical Agriculture and Forestry (School of Agricultural and Rural, School of Rural Revitalization), Hainan University, Danzhou 571737, China.

出版信息

Sensors (Basel). 2024 May 1;24(9):2905. doi: 10.3390/s24092905.

DOI:10.3390/s24092905
PMID:38733010
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11086134/
Abstract

Underwater visual detection technology is crucial for marine exploration and monitoring. Given the growing demand for accurate underwater target recognition, this study introduces an innovative architecture, YOLOv8-MU, which significantly enhances the detection accuracy. This model incorporates the large kernel block (LarK block) from UniRepLKNet to optimize the backbone network, achieving a broader receptive field without increasing the model's depth. Additionally, the integration of C2fSTR, which combines the Swin transformer with the C2f module, and the SPPFCSPC_EMA module, which blends Cross-Stage Partial Fast Spatial Pyramid Pooling (SPPFCSPC) with attention mechanisms, notably improves the detection accuracy and robustness for various biological targets. A fusion block from DAMO-YOLO further enhances the multi-scale feature extraction capabilities in the model's neck. Moreover, the adoption of the MPDIoU loss function, designed around the vertex distance, effectively addresses the challenges of localization accuracy and boundary clarity in underwater organism detection. The experimental results on the URPC2019 dataset indicate that YOLOv8-MU achieves an mAP@0.5 of 78.4%, showing an improvement of 4.0% over the original YOLOv8 model. Additionally, on the URPC2020 dataset, it achieves 80.9%, and, on the Aquarium dataset, it reaches 75.5%, surpassing other models, including YOLOv5 and YOLOv8n, thus confirming the wide applicability and generalization capabilities of our proposed improved model architecture. Furthermore, an evaluation on the improved URPC2019 dataset demonstrates leading performance (SOTA), with an mAP@0.5 of 88.1%, further verifying its superiority on this dataset. These results highlight the model's broad applicability and generalization capabilities across various underwater datasets.

摘要

水下视觉检测技术对于海洋探索和监测至关重要。鉴于对精确水下目标识别的需求不断增长,本研究引入了一种创新架构YOLOv8-MU,它显著提高了检测精度。该模型整合了来自UniRepLKNet的大内核块(LarK块)以优化骨干网络,在不增加模型深度的情况下实现了更广泛的感受野。此外,将Swin变压器与C2f模块相结合的C2fSTR以及将跨阶段部分快速空间金字塔池化(SPPFCSPC)与注意力机制相融合的SPPFCSPC_EMA模块的集成,显著提高了对各种生物目标的检测精度和鲁棒性。来自DAMO-YOLO的融合块进一步增强了模型颈部的多尺度特征提取能力。此外,采用围绕顶点距离设计的MPDIoU损失函数有效地解决了水下生物检测中定位精度和边界清晰度的挑战。在URPC2019数据集上的实验结果表明,YOLOv8-MU的mAP@0.5达到78.4%,比原始YOLOv8模型提高了4.0%。此外,在URPC2020数据集上,它达到了80.9%,在水族馆数据集上,它达到了75.5%,超过了包括YOLOv5和YOLOv8n在内的其他模型,从而证实了我们提出的改进模型架构的广泛适用性和泛化能力。此外,在改进的URPC2019数据集上的评估显示出领先性能(SOTA),mAP@0.5为88.1%,进一步验证了其在该数据集上的优越性。这些结果突出了该模型在各种水下数据集上的广泛适用性和泛化能力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/200cb812ff65/sensors-24-02905-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/5cc0e2c00761/sensors-24-02905-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/924b82a6c3aa/sensors-24-02905-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/fec552380c51/sensors-24-02905-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/b40b1e191f39/sensors-24-02905-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/7f03b86b6aa0/sensors-24-02905-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/dea198876c46/sensors-24-02905-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/5ca2bdd78f44/sensors-24-02905-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/ac74acdac164/sensors-24-02905-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/5543829cb04e/sensors-24-02905-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/342262e103dd/sensors-24-02905-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/51745554d79a/sensors-24-02905-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/dc27f7a1090d/sensors-24-02905-g012a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/f6c5032c5936/sensors-24-02905-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/38c70f05a1c9/sensors-24-02905-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/67f49b462c8d/sensors-24-02905-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/200cb812ff65/sensors-24-02905-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/5cc0e2c00761/sensors-24-02905-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/924b82a6c3aa/sensors-24-02905-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/fec552380c51/sensors-24-02905-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/b40b1e191f39/sensors-24-02905-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/7f03b86b6aa0/sensors-24-02905-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/dea198876c46/sensors-24-02905-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/5ca2bdd78f44/sensors-24-02905-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/ac74acdac164/sensors-24-02905-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/5543829cb04e/sensors-24-02905-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/342262e103dd/sensors-24-02905-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/51745554d79a/sensors-24-02905-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/dc27f7a1090d/sensors-24-02905-g012a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/f6c5032c5936/sensors-24-02905-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/38c70f05a1c9/sensors-24-02905-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/67f49b462c8d/sensors-24-02905-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7fc/11086134/200cb812ff65/sensors-24-02905-g016.jpg

相似文献

1
YOLOv8-MU: An Improved YOLOv8 Underwater Detector Based on a Large Kernel Block and a Multi-Branch Reparameterization Module.YOLOv8-MU:一种基于大内核模块和多分支重参数化模块的改进型YOLOv8水下探测器。
Sensors (Basel). 2024 May 1;24(9):2905. doi: 10.3390/s24092905.
2
FEB-YOLOv8: A multi-scale lightweight detection model for underwater object detection.FEB-YOLOv8:一种用于水下目标检测的多尺度轻量级检测模型。
PLoS One. 2024 Sep 27;19(9):e0311173. doi: 10.1371/journal.pone.0311173. eCollection 2024.
3
An Improved YOLOv5-Based Underwater Object-Detection Framework.一种改进的基于 YOLOv5 的水下目标检测框架。
Sensors (Basel). 2023 Apr 3;23(7):3693. doi: 10.3390/s23073693.
4
YOLOv8-C2f-Faster-EMA: An Improved Underwater Trash Detection Model Based on YOLOv8.YOLOv8-C2f-Faster-EMA:一种基于YOLOv8的改进型水下垃圾检测模型。
Sensors (Basel). 2024 Apr 12;24(8):2483. doi: 10.3390/s24082483.
5
YOLOv8-RMDA: Lightweight YOLOv8 Network for Early Detection of Small Target Diseases in Tea.YOLOv8-RMDA:用于茶中早期检测小目标疾病的轻量级 YOLOv8 网络。
Sensors (Basel). 2024 May 1;24(9):2896. doi: 10.3390/s24092896.
6
GFI-YOLOv8: Sika Deer Posture Recognition Target Detection Method Based on YOLOv8.GFI-YOLOv8:基于YOLOv8的梅花鹿姿态识别目标检测方法
Animals (Basel). 2024 Sep 11;14(18):2640. doi: 10.3390/ani14182640.
7
Small object detection algorithm incorporating swin transformer for tea buds.用于茶芽的融合 Swin 变换小目标检测算法。
PLoS One. 2024 Mar 21;19(3):e0299902. doi: 10.1371/journal.pone.0299902. eCollection 2024.
8
Underwater small target detection under YOLOv8-LA model.基于YOLOv8-LA模型的水下小目标检测
Sci Rep. 2024 Jul 12;14(1):16108. doi: 10.1038/s41598-024-66950-w.
9
RSE-YOLOv8: An Algorithm for Underwater Biological Target Detection.RSE-YOLOv8:一种水下生物目标检测算法。
Sensors (Basel). 2024 Sep 18;24(18):6030. doi: 10.3390/s24186030.
10
SD-YOLOv8: An Accurate Detection Model Based on Improved YOLOv8.SD-YOLOv8:一种基于改进 YOLOv8 的精确检测模型。
Sensors (Basel). 2024 Jun 4;24(11):3647. doi: 10.3390/s24113647.

引用本文的文献

1
Visualizing Plant Disease Distribution and Evaluating Model Performance for Deep Learning Classification with YOLOv8.利用YOLOv8可视化植物病害分布并评估深度学习分类模型性能
Pathogens. 2024 Nov 22;13(12):1032. doi: 10.3390/pathogens13121032.
2
Methodology for the correction of a CBCT volume from the skull to the natural head position.将锥形束计算机断层扫描(CBCT)容积从颅骨位置校正到自然头部位置的方法。
MethodsX. 2024 Nov 27;13:103073. doi: 10.1016/j.mex.2024.103073. eCollection 2024 Dec.
3
Dense Pedestrian Detection Based on GR-YOLO.

本文引用的文献

1
Cattle Body Detection Based on YOLOv5-EMA for Precision Livestock Farming.基于YOLOv5-EMA的牛体检测用于精准畜牧养殖
Animals (Basel). 2023 Nov 15;13(22):3535. doi: 10.3390/ani13223535.
2
Tea-YOLOv8s: A Tea Bud Detection Model Based on Deep Learning and Computer Vision.茶YOLOv8s:一种基于深度学习和计算机视觉的茶芽检测模型。
Sensors (Basel). 2023 Jul 21;23(14):6576. doi: 10.3390/s23146576.
3
Domain Adaptation for Underwater Image Enhancement.水下图像增强的域适应
基于GR-YOLO的密集行人检测
Sensors (Basel). 2024 Jul 22;24(14):4747. doi: 10.3390/s24144747.
IEEE Trans Image Process. 2023;32:1442-1457. doi: 10.1109/TIP.2023.3244647. Epub 2023 Mar 1.
4
An Underwater Human-Robot Interaction Using a Visual-Textual Model for Autonomous Underwater Vehicles.基于视觉-文本模型的自主水下机器人水下人机交互
Sensors (Basel). 2022 Dec 24;23(1):197. doi: 10.3390/s23010197.
5
A Big Coal Block Alarm Detection Method for Scraper Conveyor Based on YOLO-BS.基于 YOLO-BS 的刮板输送机大煤块报警检测方法。
Sensors (Basel). 2022 Nov 22;22(23):9052. doi: 10.3390/s22239052.
6
Projecting future changes in distributions of small-scale pelagic fisheries of the southern Colombian Pacific Ocean.预测哥伦比亚太平洋沿岸南部小型中上层渔业分布的未来变化。
Heliyon. 2022 Feb 18;8(2):e08975. doi: 10.1016/j.heliyon.2022.e08975. eCollection 2022 Feb.
7
Underwater image enhancement using adaptive color restoration and dehazing.基于自适应色彩恢复与去雾的水下图像增强
Opt Express. 2022 Feb 14;30(4):6216-6235. doi: 10.1364/OE.449930.
8
Learning to Match Anchors for Visual Object Detection.学习用于视觉目标检测的锚匹配。
IEEE Trans Pattern Anal Mach Intell. 2022 Jun;44(6):3096-3109. doi: 10.1109/TPAMI.2021.3050494. Epub 2022 May 5.
9
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.
10
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.