• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种用于PET/CT图像中肺肿瘤的自适应轻量级YOLOv5检测模型。

An adaptive and lightweight YOLOv5 detection model for lung tumor in PET/CT images.

作者信息

Zhou Tao, Ye Xinyu, Lu Huiling, Guo Yujie, Wang Hongxia, Liu Yang

机构信息

School of Computer Science and Engineering, North Minzu University, Yinchuan, 750021, China.

Key Laboratory of Image and Graphics Intelligent Processing of State Ethnic Affairs Commission, North Minzu University, Yinchuan, 750021, China.

出版信息

Sci Rep. 2024 Dec 28;14(1):30719. doi: 10.1038/s41598-024-79786-1.

DOI:10.1038/s41598-024-79786-1
PMID:39730429
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11681088/
Abstract

Multi-modal medical images are important in tumor lesion detection. However, the existing detection models only use single-modal to detect lesions, a multi-modal semantic correlation is not enough to consider and lacks ability to express the shape, size, and contrast degree features of lesions. A Cross Modal YOLOv5 model (CMYOLOv5) is proposed. Firstly, there are two networks, auxiliary network is consisted by dual-branch structure to extract semantic information from PET and CT, backbone network is consisted by YOLOv5 to extract semantic information from PET/CT. Secondly, Cross-modal Features Fusion (CFF) is designed in auxiliary network to fuse PET functional information and CT anatomical information. Self-Adaptive Attention Fusion (AAF) is designed in backbone network to fuse and enhance three-modal complementary information. Thirdly, Self-Adaptive Transformer (SAT) is designed in feature enhance neck. Using Transformer with deformable attention mechanism to focus on lung tumor region. Using MLP with channel attention mechanism to enhance features representation ability of lung tumor region. Finally, Reparameter Residual Block (RRB) and Reparameter Convolution operation (RC) are designed to fully learn richer PET, CT and PET/CT feature. Comparative experiments are conducted on clinical lung tumor PET/CT multi-modality dataset, the effectiveness of CMYOLOv5 is verified by Precision, Recall, mAP, F1, FPS, and training time, experimental results are 97.16%, 96.41%, 97.18%, 96.78%, 96.37 and 3912 s. CMYOLOv5 has high precision in the detection of irregular lung tumors, which is superior to the existing advanced methods.

摘要

多模态医学图像在肿瘤病变检测中具有重要意义。然而,现有的检测模型仅使用单模态来检测病变,对多模态语义相关性的考虑不足,并且缺乏表达病变形状、大小和对比度特征的能力。为此提出了一种跨模态YOLOv5模型(CMYOLOv5)。首先,该模型有两个网络,辅助网络由双分支结构组成,用于从PET和CT中提取语义信息,主干网络由YOLOv5组成,用于从PET/CT中提取语义信息。其次,在辅助网络中设计了跨模态特征融合(CFF),以融合PET功能信息和CT解剖信息。在主干网络中设计了自适应注意力融合(AAF),以融合和增强三模态互补信息。第三,在特征增强颈部设计了自适应Transformer(SAT)。使用具有可变形注意力机制的Transformer聚焦于肺肿瘤区域。使用具有通道注意力机制的MLP增强肺肿瘤区域的特征表示能力。最后,设计了重参数化残差块(RRB)和重参数化卷积操作(RC),以充分学习更丰富的PET、CT和PET/CT特征。在临床肺肿瘤PET/CT多模态数据集上进行了对比实验,通过精度、召回率、平均精度均值(mAP)、F1值、每秒帧数(FPS)和训练时间验证了CMYOLOv5的有效性,实验结果分别为97.16%、96.41%、97.18%、96.78%、96.37和3912秒。CMYOLOv5在不规则肺肿瘤检测中具有较高的精度,优于现有的先进方法。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/5250bc389d36/41598_2024_79786_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/45bb4a3bf76e/41598_2024_79786_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/9be7cdf4b1be/41598_2024_79786_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/4f4fd6a9f8ab/41598_2024_79786_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/765b89605402/41598_2024_79786_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/6efcb3ae0f7f/41598_2024_79786_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/c9f9ff707517/41598_2024_79786_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/f7107e068c03/41598_2024_79786_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/d18f18d0d2b8/41598_2024_79786_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/340d4fde2e74/41598_2024_79786_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/40dceb3eef57/41598_2024_79786_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/5d1207deb41f/41598_2024_79786_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/d99a181c2921/41598_2024_79786_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/2bdf2cf6972e/41598_2024_79786_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/0df18799903a/41598_2024_79786_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/efa8de73dd81/41598_2024_79786_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/5250bc389d36/41598_2024_79786_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/45bb4a3bf76e/41598_2024_79786_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/9be7cdf4b1be/41598_2024_79786_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/4f4fd6a9f8ab/41598_2024_79786_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/765b89605402/41598_2024_79786_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/6efcb3ae0f7f/41598_2024_79786_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/c9f9ff707517/41598_2024_79786_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/f7107e068c03/41598_2024_79786_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/d18f18d0d2b8/41598_2024_79786_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/340d4fde2e74/41598_2024_79786_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/40dceb3eef57/41598_2024_79786_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/5d1207deb41f/41598_2024_79786_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/d99a181c2921/41598_2024_79786_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/2bdf2cf6972e/41598_2024_79786_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/0df18799903a/41598_2024_79786_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/efa8de73dd81/41598_2024_79786_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5269/11681088/5250bc389d36/41598_2024_79786_Fig16_HTML.jpg

相似文献

1
An adaptive and lightweight YOLOv5 detection model for lung tumor in PET/CT images.一种用于PET/CT图像中肺肿瘤的自适应轻量级YOLOv5检测模型。
Sci Rep. 2024 Dec 28;14(1):30719. doi: 10.1038/s41598-024-79786-1.
2
CCGL-YOLOV5:A cross-modal cross-scale global-local attention YOLOV5 lung tumor detection model.CCGL-YOLOV5:一种跨模态跨尺度全局-局部注意力 YOLOV5 肺肿瘤检测模型。
Comput Biol Med. 2023 Oct;165:107387. doi: 10.1016/j.compbiomed.2023.107387. Epub 2023 Aug 28.
3
SwinCross: Cross-modal Swin transformer for head-and-neck tumor segmentation in PET/CT images.SwinCross:用于 PET/CT 图像中头颈部肿瘤分割的跨模态 Swin 变换器。
Med Phys. 2024 Mar;51(3):2096-2107. doi: 10.1002/mp.16703. Epub 2023 Sep 30.
4
MFCNet: A multi-modal fusion and calibration networks for 3D pancreas tumor segmentation on PET-CT images.MFCNet:一种用于PET-CT图像上3D胰腺肿瘤分割的多模态融合与校准网络。
Comput Biol Med. 2023 Mar;155:106657. doi: 10.1016/j.compbiomed.2023.106657. Epub 2023 Feb 10.
5
A transformer-guided cross-modality adaptive feature fusion framework for esophageal gross tumor volume segmentation.基于Transformer 的跨模态自适应特征融合框架用于食管大体肿瘤体积分割。
Comput Methods Programs Biomed. 2024 Jun;251:108216. doi: 10.1016/j.cmpb.2024.108216. Epub 2024 May 11.
6
DMCT-Net: dual modules convolution transformer network for head and neck tumor segmentation in PET/CT.DMCT-Net:用于 PET/CT 中头颈部肿瘤分割的双模块卷积变换网络。
Phys Med Biol. 2023 May 22;68(11). doi: 10.1088/1361-6560/acd29f.
7
MSRA-Net: Tumor segmentation network based on Multi-scale Residual Attention.MSRA-Net:基于多尺度残差注意力的肿瘤分割网络。
Comput Biol Med. 2023 May;158:106818. doi: 10.1016/j.compbiomed.2023.106818. Epub 2023 Mar 22.
8
[Pulmonary PET /CT image instance segmentation based on dense interactive feature fusion Mask RCNN].基于密集交互式特征融合Mask RCNN的肺部PET/CT图像实例分割
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi. 2024 Jun 25;41(3):527-534. doi: 10.7507/1001-5515.202309026.
9
DGCBG-Net: A dual-branch network with global cross-modal interaction and boundary guidance for tumor segmentation in PET/CT images.DGCBG-Net:一种具有全局跨模态交互和边界引导的双分支网络,用于 PET/CT 图像中的肿瘤分割。
Comput Methods Programs Biomed. 2024 Jun;250:108125. doi: 10.1016/j.cmpb.2024.108125. Epub 2024 Mar 20.
10
GMRE-iUnet: Isomorphic Unet fusion model for PET and CT lung tumor images.GMRE-iUnet:用于 PET 和 CT 肺肿瘤图像的同态 U-Net 融合模型。
Comput Biol Med. 2023 Nov;166:107514. doi: 10.1016/j.compbiomed.2023.107514. Epub 2023 Sep 28.

本文引用的文献

1
CCGL-YOLOV5:A cross-modal cross-scale global-local attention YOLOV5 lung tumor detection model.CCGL-YOLOV5:一种跨模态跨尺度全局-局部注意力 YOLOV5 肺肿瘤检测模型。
Comput Biol Med. 2023 Oct;165:107387. doi: 10.1016/j.compbiomed.2023.107387. Epub 2023 Aug 28.
2
An improved faster R-CNN algorithm for assisted detection of lung nodules.一种改进的更快的 R-CNN 算法,用于辅助肺结节检测。
Comput Biol Med. 2023 Feb;153:106470. doi: 10.1016/j.compbiomed.2022.106470. Epub 2022 Dec 28.
3
Two-stage lung nodule detection framework using enhanced UNet and convolutional LSTM networks in CT images.
基于增强型 UNet 和卷积 LSTM 网络的 CT 图像两阶段肺结节检测框架。
Comput Biol Med. 2022 Oct;149:106059. doi: 10.1016/j.compbiomed.2022.106059. Epub 2022 Sep 3.
4
Computer-aided detection and segmentation of malignant melanoma lesions on whole-body F-FDG PET/CT using an interpretable deep learning approach.基于可解释深度学习方法的全身 F-FDG PET/CT 中恶性黑色素瘤病变的计算机辅助检测和分割。
Comput Methods Programs Biomed. 2022 Jun;221:106902. doi: 10.1016/j.cmpb.2022.106902. Epub 2022 May 22.
5
Deep learning-based multimodal image analysis for cervical cancer detection.基于深度学习的宫颈癌多模态图像分析。
Methods. 2022 Sep;205:46-52. doi: 10.1016/j.ymeth.2022.05.004. Epub 2022 May 20.
6
Dense Convolutional Network and Its Application in Medical Image Analysis.密集卷积网络及其在医学图像分析中的应用。
Biomed Res Int. 2022 Apr 25;2022:2384830. doi: 10.1155/2022/2384830. eCollection 2022.
7
One-stage pulmonary nodule detection using 3-D DCNN with feature fusion and attention mechanism in CT image.基于 CT 图像三维 DCNN 特征融合与注意力机制的肺结节一站式检测
Comput Methods Programs Biomed. 2022 Jun;220:106786. doi: 10.1016/j.cmpb.2022.106786. Epub 2022 Apr 4.
8
The ensemble deep learning model for novel COVID-19 on CT images.用于新型冠状病毒肺炎CT图像的集成深度学习模型。
Appl Soft Comput. 2021 Jan;98:106885. doi: 10.1016/j.asoc.2020.106885. Epub 2020 Nov 6.
9
Deep Learning for Variational Multimodality Tumor Segmentation in PET/CT.用于PET/CT中变分多模态肿瘤分割的深度学习
Neurocomputing (Amst). 2020 Jun 7;392:277-295. doi: 10.1016/j.neucom.2018.10.099. Epub 2019 Apr 24.
10
Multi-model Ensemble Learning Architecture Based on 3D CNN for Lung Nodule Malignancy Suspiciousness Classification.基于 3DCNN 的多模态集成学习架构用于肺结节良恶性可疑度分类。
J Digit Imaging. 2020 Oct;33(5):1242-1256. doi: 10.1007/s10278-020-00372-8.