• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

VariMix:用于可解释医学图像分类的基于多样性引导的数据混合框架。

VariMix: A variety-guided data mixing framework for explainable medical image classifications.

作者信息

Xiong Xiangyu, Sun Yue, Liu Xiaohong, Ke Wei, Lam Chan-Tong, Gao Qinquan, Tong Tong, Li Shuo, Tan Tao

机构信息

Faculty of Applied Sciences, Macao Polytechnic University, Macao, 999078, China.

John Hopcroft Center (JHC) for Computer Science, Shanghai Jiao Tong University, Shanghai, 200240, China.

出版信息

Comput Methods Programs Biomed. 2025 Nov;271:109016. doi: 10.1016/j.cmpb.2025.109016. Epub 2025 Aug 16.

DOI:10.1016/j.cmpb.2025.109016
PMID:40840261
Abstract

BACKGROUND AND OBJECTIVE

Modern deep neural networks are highly over-parameterized, necessitating the use of data augmentation techniques to prevent overfitting and enhance generalization. Generative adversarial networks (GANs) are popular for synthesizing visually realistic images. However, these synthetic images often lack diversity and may have ambiguous class labels. Recent data mixing strategies address some of these issues by mixing image labels based on salient regions. Since the main diagnostic information is not always contained within the salient regions, we aim to address the resulting label mismatches in medical image classifications.

METHODS

We propose a variety-guided data mixing framework (VariMix), which exploits an absolute difference map (ADM) to address the label mismatch problems of mixed medical images. VariMix generates ADM using the image-to-image (I2I) GAN across multiple classes and allows for bidirectional mixing operations between the training samples.

RESULTS

The proposed VariMix achieves the highest accuracy of 99.30% and 94.60% with a SwinT V2 classifier on a Chest X-ray (CXR) dataset and a Retinal dataset, respectively. It also achieves the highest accuracy of 87.73%, 99.28%, 95.13%, and 95.81% with a ConvNeXt classifier on a Breast Ultrasound (US) dataset, a CXR dataset, a Retinal dataset, and a Maternal-Fetal US dataset, respectively. Furthermore, the medical expert evaluation on generated images shows the great potential of our proposed I2I GAN in improving the accuracy of medical image classifications.

CONCLUSIONS

Extensive experiments demonstrate the superiority of VariMix compared with the existing GAN- and Mixup-based methods on four public datasets using Swin Transformer V2 and ConvNeXt architectures. Furthermore, by projecting the source image to the hyperplanes of the classifiers, the proposed I2I GAN can generate hyperplane difference maps between the source image and the hyperplane image, demonstrating its ability to interpret medical image classifications. The source code is provided in https://github.com/yXiangXiong/VariMix.

摘要

背景与目的

现代深度神经网络具有高度的过参数化特性,因此需要使用数据增强技术来防止过拟合并提高泛化能力。生成对抗网络(GAN)在合成视觉上逼真的图像方面很受欢迎。然而,这些合成图像往往缺乏多样性,并且可能具有模糊的类别标签。最近的数据混合策略通过基于显著区域混合图像标签来解决其中一些问题。由于主要诊断信息并不总是包含在显著区域内,我们旨在解决医学图像分类中由此产生的标签不匹配问题。

方法

我们提出了一种多类别引导的数据混合框架(VariMix),该框架利用绝对差异图(ADM)来解决混合医学图像的标签不匹配问题。VariMix使用跨多个类别的图像到图像(I2I)GAN生成ADM,并允许在训练样本之间进行双向混合操作。

结果

所提出的VariMix在胸部X光(CXR)数据集和视网膜数据集上,使用SwinT V2分类器分别达到了99.30%和94.60%的最高准确率。在乳腺超声(US)数据集、CXR数据集、视网膜数据集和母胎US数据集上,使用ConvNeXt分类器分别达到了87.73%、99.28%、95.13%和95.81%的最高准确率。此外,医学专家对生成图像的评估表明,我们提出的I2I GAN在提高医学图像分类准确率方面具有巨大潜力。

结论

大量实验表明,在使用Swin Transformer V2和ConvNeXt架构的四个公共数据集上,VariMix与现有的基于GAN和Mixup的方法相比具有优越性。此外,通过将源图像投影到分类器的超平面上,所提出的I2I GAN可以生成源图像与超平面图像之间的超平面差异图,证明了其解释医学图像分类的能力。源代码可在https://github.com/yXiangXiong/VariMix获取。

相似文献

1
VariMix: A variety-guided data mixing framework for explainable medical image classifications.VariMix:用于可解释医学图像分类的基于多样性引导的数据混合框架。
Comput Methods Programs Biomed. 2025 Nov;271:109016. doi: 10.1016/j.cmpb.2025.109016. Epub 2025 Aug 16.
2
A medical image classification method based on self-regularized adversarial learning.基于自正则化对抗学习的医学图像分类方法。
Med Phys. 2024 Nov;51(11):8232-8246. doi: 10.1002/mp.17320. Epub 2024 Jul 30.
3
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
4
CXR-MultiTaskNet a unified deep learning framework for joint disease localization and classification in chest radiographs.CXR-MultiTaskNet:一种用于胸部X光片中疾病联合定位与分类的统一深度学习框架。
Sci Rep. 2025 Aug 31;15(1):32022. doi: 10.1038/s41598-025-16669-z.
5
Noise-aware system generative model (NASGM): positron emission tomography (PET) image simulation framework with observer validation studies.噪声感知系统生成模型(NASGM):用于正电子发射断层扫描(PET)图像模拟框架及观察者验证研究。
Med Phys. 2025 Jul;52(7):e17962. doi: 10.1002/mp.17962.
6
Leveraging a foundation model zoo for cell similarity search in oncological microscopy across devices.利用基础模型库进行跨设备肿瘤显微镜检查中的细胞相似性搜索。
Front Oncol. 2025 Jun 18;15:1480384. doi: 10.3389/fonc.2025.1480384. eCollection 2025.
7
Generalizable diagnosis of chest radiographs through attention-guided decomposition of images utilizing self-consistency loss.利用自一致性损失引导图像分解进行可推广的胸片诊断。
Comput Biol Med. 2024 Sep;180:108922. doi: 10.1016/j.compbiomed.2024.108922. Epub 2024 Jul 31.
8
A segment anything model-guided and match-based semi-supervised segmentation framework for medical imaging.一种用于医学成像的基于段式分割模型引导和匹配的半监督分割框架。
Med Phys. 2025 Mar 29. doi: 10.1002/mp.17785.
9
An open-source deep learning framework for respiratory motion monitoring and volumetric imaging during radiation therapy.一种用于放射治疗期间呼吸运动监测和容积成像的开源深度学习框架。
Med Phys. 2025 Jul;52(7):e18015. doi: 10.1002/mp.18015.
10
A comprehensive comparative study of generative adversarial network architectures for synthetic computed tomography generation in the abdomen.用于腹部合成计算机断层扫描生成的生成对抗网络架构的全面比较研究。
Med Phys. 2025 Aug;52(8):e18038. doi: 10.1002/mp.18038.