• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于处理医学图像中噪声标签和加权冗余的全局上下文和局部交互中的部分注意力机制

Partial Attention in Global Context and Local Interaction for Addressing Noisy Labels and Weighted Redundancies on Medical Images.

作者信息

Nguyen Minh Tai Pham, Phan Tran Minh Khue, Nakano Tadashi, Tran Thi Hong, Nguyen Quoc Duy Nam

机构信息

Faculty of Advanced Program, Ho Chi Minh City Open University, Ho Chi Minh City 700000, Vietnam.

Faculty of Information Technology, Ho Chi Minh City Open University, Ho Chi Minh City 700000, Vietnam.

出版信息

Sensors (Basel). 2024 Dec 30;25(1):163. doi: 10.3390/s25010163.

DOI:10.3390/s25010163
PMID:39796954
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11722591/
Abstract

Recently, the application of deep neural networks to detect anomalies on medical images has been facing the appearance of noisy labels, including overlapping objects and similar classes. Therefore, this study aims to address this challenge by proposing a unique attention module that can assist deep neural networks in focusing on important object features in noisy medical image conditions. This module integrates global context modeling to create long-range dependencies and local interactions to enable channel attention ability by using 1D convolution that not only performs well with noisy labels but also consumes significantly less resources without any dimensionality reduction. The module is then named Global Context and Local Interaction (GCLI). We have further experimented and proposed a partial attention strategy for the proposed GCLI module, aiming to efficiently reduce weighted redundancies. This strategy utilizes a subset of channels for GCLI to produce attention weights instead of considering every single channel. As a result, this strategy can greatly reduce the risk of introducing weighted redundancies caused by modeling global context. For classification, our proposed method is able to assist ResNet34 in achieving up to 82.5% accuracy on the Chaoyang test set, which is the highest figure among the other SOTA attention modules without using any processing filter to reduce the effect of noisy labels. For object detection, the GCLI is able to boost the capability of YOLOv8 up to 52.1% mAP50 on the GRAZPEDWRI-DX test set, demonstrating the highest performance among other attention modules and ranking second in the mAP50 metric on the VinDR-CXR test set. In terms of model complexity, our proposed GCLI module can consume fewer extra parameters up to 225 times and has inference speed faster than 30% compared to the other attention modules.

摘要

最近,将深度神经网络应用于医学图像异常检测面临着噪声标签的出现,包括重叠对象和相似类别。因此,本研究旨在通过提出一种独特的注意力模块来应对这一挑战,该模块可以帮助深度神经网络在有噪声的医学图像条件下专注于重要的对象特征。该模块集成了全局上下文建模以创建长程依赖关系,并通过使用1D卷积实现局部交互以具备通道注意力能力,这种1D卷积不仅在处理噪声标签时表现良好,而且在不进行任何降维的情况下消耗的资源显著减少。该模块随后被命名为全局上下文与局部交互(GCLI)。我们进一步进行了实验,并为所提出的GCLI模块提出了一种部分注意力策略,旨在有效减少加权冗余。该策略利用GCLI的通道子集来生成注意力权重,而不是考虑每一个通道。结果,该策略可以大大降低因对全局上下文建模而引入加权冗余的风险。对于分类,我们提出的方法能够帮助ResNet34在朝阳测试集上达到高达82.5%的准确率,这是在不使用任何处理滤波器来减少噪声标签影响的情况下,其他最先进注意力模块中的最高数字。对于目标检测,GCLI能够在GRAZPEDWRI-DX测试集上将YOLOv8的能力提升至52.1% mAP50,在其他注意力模块中表现出最高性能,在VinDR-CXR测试集的mAP50指标中排名第二。在模型复杂度方面,我们提出的GCLI模块与其他注意力模块相比,可以消耗少至225倍的额外参数,推理速度快30%以上。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/b8af815d3e9e/sensors-25-00163-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/c7bfa53e9e38/sensors-25-00163-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/908bfdfafda7/sensors-25-00163-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/e38a2156a559/sensors-25-00163-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/a27a4976ccd2/sensors-25-00163-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/6cc5f077243b/sensors-25-00163-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/d4106f0400b0/sensors-25-00163-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/bf8c62270e56/sensors-25-00163-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/f5a57eb01c0d/sensors-25-00163-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/83827dbd7a97/sensors-25-00163-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/ca79d8bc1ce5/sensors-25-00163-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/1b82f8822258/sensors-25-00163-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/122719d092b1/sensors-25-00163-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/68a07436323d/sensors-25-00163-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/60127a300bfa/sensors-25-00163-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/9cad009bbb44/sensors-25-00163-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/76b30567ac52/sensors-25-00163-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/efbe9f49e948/sensors-25-00163-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/b8af815d3e9e/sensors-25-00163-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/c7bfa53e9e38/sensors-25-00163-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/908bfdfafda7/sensors-25-00163-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/e38a2156a559/sensors-25-00163-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/a27a4976ccd2/sensors-25-00163-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/6cc5f077243b/sensors-25-00163-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/d4106f0400b0/sensors-25-00163-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/bf8c62270e56/sensors-25-00163-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/f5a57eb01c0d/sensors-25-00163-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/83827dbd7a97/sensors-25-00163-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/ca79d8bc1ce5/sensors-25-00163-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/1b82f8822258/sensors-25-00163-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/122719d092b1/sensors-25-00163-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/68a07436323d/sensors-25-00163-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/60127a300bfa/sensors-25-00163-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/9cad009bbb44/sensors-25-00163-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/76b30567ac52/sensors-25-00163-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/efbe9f49e948/sensors-25-00163-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f215/11722591/b8af815d3e9e/sensors-25-00163-g018.jpg

相似文献

1
Partial Attention in Global Context and Local Interaction for Addressing Noisy Labels and Weighted Redundancies on Medical Images.用于处理医学图像中噪声标签和加权冗余的全局上下文和局部交互中的部分注意力机制
Sensors (Basel). 2024 Dec 30;25(1):163. doi: 10.3390/s25010163.
2
Brain tumor segmentation and detection in MRI using convolutional neural networks and VGG16.使用卷积神经网络和VGG16在磁共振成像(MRI)中进行脑肿瘤分割与检测
Cancer Biomark. 2025 Mar;42(3):18758592241311184. doi: 10.1177/18758592241311184. Epub 2025 Apr 4.
3
Sample self-selection using dual teacher networks for pathological image classification with noisy labels.使用双教师网络进行带噪标签的病理图像分类的样本自选择。
Comput Biol Med. 2024 May;174:108489. doi: 10.1016/j.compbiomed.2024.108489. Epub 2024 Apr 16.
4
Suppressing label noise in medical image classification using mixup attention and self-supervised learning.利用混叠注意力和自监督学习抑制医学图像分类中的标签噪声。
Phys Med Biol. 2024 May 8;69(10). doi: 10.1088/1361-6560/ad4083.
5
Object recognition in medical images via anatomy-guided deep learning.通过解剖学引导的深度学习实现医学图像中的目标识别。
Med Image Anal. 2022 Oct;81:102527. doi: 10.1016/j.media.2022.102527. Epub 2022 Jun 25.
6
GlobalSR: Global context network for single image super-resolution via deformable convolution attention and fast Fourier convolution.GlobalSR:基于可变形卷积注意力和快速傅里叶卷积的单图像超分辨率全局上下文网络。
Neural Netw. 2024 Dec;180:106686. doi: 10.1016/j.neunet.2024.106686. Epub 2024 Aug 31.
7
Robust co-teaching learning with consistency-based noisy label correction for medical image classification.用于医学图像分类的基于一致性的噪声标签校正的稳健协同教学学习
Int J Comput Assist Radiol Surg. 2023 Apr;18(4):675-683. doi: 10.1007/s11548-022-02799-6. Epub 2022 Nov 27.
8
IEA-Net: Internal and External Dual-Attention Medical Segmentation Network with High-Performance Convolutional Blocks.IEA-Net:具有高性能卷积块的内外双注意力医学分割网络
J Imaging Inform Med. 2025 Feb;38(1):602-614. doi: 10.1007/s10278-024-01217-4. Epub 2024 Aug 6.
9
CFANet: Context Feature Fusion and Attention Mechanism Based Network for Small Target Segmentation in Medical Images.CFANet:基于上下文特征融合和注意力机制的医学图像小目标分割网络。
Sensors (Basel). 2023 Oct 26;23(21):8739. doi: 10.3390/s23218739.
10
GC-Net: Global context network for medical image segmentation.GC-Net:用于医学图像分割的全局上下文网络。
Comput Methods Programs Biomed. 2020 Jul;190:105121. doi: 10.1016/j.cmpb.2019.105121. Epub 2019 Oct 4.

本文引用的文献

1
Self-supervised learning for medical image classification: a systematic review and implementation guidelines.用于医学图像分类的自监督学习:系统综述与实施指南
NPJ Digit Med. 2023 Apr 26;6(1):74. doi: 10.1038/s41746-023-00811-0.
2
MedViT: A robust vision transformer for generalized medical image classification.MedViT:一种用于广义医学图像分类的鲁棒视觉Transformer。
Comput Biol Med. 2023 May;157:106791. doi: 10.1016/j.compbiomed.2023.106791. Epub 2023 Mar 14.
3
Lesion detection of chest X-Ray based on scalable attention residual CNN.
基于可扩展注意力残差卷积神经网络的胸部 X 射线病灶检测。
Math Biosci Eng. 2023 Jan;20(2):1730-1749. doi: 10.3934/mbe.2023079. Epub 2022 Nov 4.
4
Why does demand for medical imaging keep rising?为何医学成像的需求持续上升?
BMJ. 2022 Nov 2;379:o2614. doi: 10.1136/bmj.o2614.
5
VinDr-CXR: An open dataset of chest X-rays with radiologist's annotations.VinDr-CXR:一个带有放射科医生标注的胸部 X 光数据集。
Sci Data. 2022 Jul 20;9(1):429. doi: 10.1038/s41597-022-01498-w.
6
Vision transformer and explainable transfer learning models for auto detection of kidney cyst, stone and tumor from CT-radiography.基于视觉Transformer和可解释迁移学习模型的 CT 放射图像肾囊肿、结石和肿瘤自动检测
Sci Rep. 2022 Jul 6;12(1):11440. doi: 10.1038/s41598-022-15634-4.
7
Explainable Vision Transformers and Radiomics for COVID-19 Detection in Chest X-rays.用于胸部X光片中COVID-19检测的可解释视觉Transformer与放射组学
J Clin Med. 2022 May 26;11(11):3013. doi: 10.3390/jcm11113013.
8
A pediatric wrist trauma X-ray dataset (GRAZPEDWRI-DX) for machine learning.儿科腕部创伤 X 射线数据集(GRAZPEDWRI-DX)用于机器学习。
Sci Data. 2022 May 20;9(1):222. doi: 10.1038/s41597-022-01328-z.
9
Pooling in convolutional neural networks for medical image analysis: a survey and an empirical study.用于医学图像分析的卷积神经网络池化:一项综述与实证研究
Neural Comput Appl. 2022;34(7):5321-5347. doi: 10.1007/s00521-022-06953-8. Epub 2022 Feb 1.
10
Robust Medical Image Classification From Noisy Labeled Data With Global and Local Representation Guided Co-Training.基于全局和局部表示引导协同训练的含噪标记数据的稳健医学图像分类。
IEEE Trans Med Imaging. 2022 Jun;41(6):1371-1382. doi: 10.1109/TMI.2021.3140140. Epub 2022 Jun 1.