• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

在磁共振图像分类中评估预训练对解释性能的影响。

Benchmarking the influence of pre-training on explanation performance in MR image classification.

作者信息

Oliveira Marta, Wilming Rick, Clark Benedict, Budding Céline, Eitel Fabian, Ritter Kerstin, Haufe Stefan

机构信息

Division 8.44, Physikalisch-Technische Bundesanstalt, Berlin, Germany.

Computer Science Department, Technische Universität Berlin, Berlin, Germany.

出版信息

Front Artif Intell. 2024 Feb 26;7:1330919. doi: 10.3389/frai.2024.1330919. eCollection 2024.

DOI:10.3389/frai.2024.1330919
PMID:38469161
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10925627/
Abstract

Convolutional Neural Networks (CNNs) are frequently and successfully used in medical prediction tasks. They are often used in combination with transfer learning, leading to improved performance when training data for the task are scarce. The resulting models are highly complex and typically do not provide any insight into their predictive mechanisms, motivating the field of "explainable" artificial intelligence (XAI). However, previous studies have rarely quantitatively evaluated the "explanation performance" of XAI methods against ground-truth data, and transfer learning and its influence on objective measures of explanation performance has not been investigated. Here, we propose a benchmark dataset that allows for quantifying explanation performance in a realistic magnetic resonance imaging (MRI) classification task. We employ this benchmark to understand the influence of transfer learning on the quality of explanations. Experimental results show that popular XAI methods applied to the same underlying model differ vastly in performance, even when considering only correctly classified examples. We further observe that explanation performance strongly depends on the task used for pre-training and the number of CNN layers pre-trained. These results hold after correcting for a substantial correlation between explanation and classification performance.

摘要

卷积神经网络(CNN)在医学预测任务中被频繁且成功地使用。它们经常与迁移学习结合使用,当该任务的训练数据稀缺时,能提高性能。由此产生的模型高度复杂,通常无法对其预测机制提供任何见解,这推动了“可解释”人工智能(XAI)领域的发展。然而,以往的研究很少针对真实数据定量评估XAI方法的“解释性能”,并且尚未研究迁移学习及其对解释性能客观指标的影响。在此,我们提出了一个基准数据集,该数据集能够在实际的磁共振成像(MRI)分类任务中量化解释性能。我们利用这个基准来了解迁移学习对解释质量的影响。实验结果表明,即使仅考虑正确分类的示例,应用于相同基础模型的流行XAI方法在性能上也存在巨大差异。我们进一步观察到,解释性能强烈依赖于用于预训练的任务以及预训练的CNN层数。在纠正了解释与分类性能之间的显著相关性之后,这些结果依然成立。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/72c5/10925627/b9c8a78b7b93/frai-07-1330919-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/72c5/10925627/606fae28c60e/frai-07-1330919-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/72c5/10925627/ba64c7f2b2bb/frai-07-1330919-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/72c5/10925627/d13232b66c32/frai-07-1330919-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/72c5/10925627/b9c8a78b7b93/frai-07-1330919-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/72c5/10925627/606fae28c60e/frai-07-1330919-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/72c5/10925627/ba64c7f2b2bb/frai-07-1330919-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/72c5/10925627/d13232b66c32/frai-07-1330919-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/72c5/10925627/b9c8a78b7b93/frai-07-1330919-g0004.jpg

相似文献

1
Benchmarking the influence of pre-training on explanation performance in MR image classification.在磁共振图像分类中评估预训练对解释性能的影响。
Front Artif Intell. 2024 Feb 26;7:1330919. doi: 10.3389/frai.2024.1330919. eCollection 2024.
2
XAI-TRIS: non-linear image benchmarks to quantify false positive post-hoc attribution of feature importance.XAI-TRIS:用于量化特征重要性的事后误报归因的非线性图像基准。
Mach Learn. 2024;113(9):6871-6910. doi: 10.1007/s10994-024-06574-3. Epub 2024 Jul 16.
3
Essential properties and explanation effectiveness of explainable artificial intelligence in healthcare: A systematic review.可解释人工智能在医疗保健中的基本属性和解释效果:一项系统综述。
Heliyon. 2023 May 8;9(5):e16110. doi: 10.1016/j.heliyon.2023.e16110. eCollection 2023 May.
4
Image Embeddings Extracted from CNNs Outperform Other Transfer Learning Approaches in Classification of Chest Radiographs.从卷积神经网络(CNNs)提取的图像嵌入在胸部X光片分类中优于其他迁移学习方法。
Diagnostics (Basel). 2022 Aug 28;12(9):2084. doi: 10.3390/diagnostics12092084.
5
Neuro-XAI: Explainable deep learning framework based on deeplabV3+ and bayesian optimization for segmentation and classification of brain tumor in MRI scans.Neuro-XAI:基于deeplabV3+和贝叶斯优化的可解释深度学习框架,用于磁共振成像扫描中脑肿瘤的分割和分类。
J Neurosci Methods. 2024 Oct;410:110247. doi: 10.1016/j.jneumeth.2024.110247. Epub 2024 Aug 10.
6
Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning.用于计算机辅助检测的深度卷积神经网络:卷积神经网络架构、数据集特征与迁移学习
IEEE Trans Med Imaging. 2016 May;35(5):1285-98. doi: 10.1109/TMI.2016.2528162. Epub 2016 Feb 11.
7
Recent Advances in Explainable Artificial Intelligence for Magnetic Resonance Imaging.用于磁共振成像的可解释人工智能的最新进展
Diagnostics (Basel). 2023 Apr 27;13(9):1571. doi: 10.3390/diagnostics13091571.
8
Ground-truth-free deep learning for artefacts reduction in 2D radial cardiac cine MRI using a synthetically generated dataset.基于合成数据集的 2D 径向心脏电影 MRI 伪影减少的无真值深度学习。
Phys Med Biol. 2021 Apr 23;66(9). doi: 10.1088/1361-6560/abf278.
9
Quantitative evaluation of Saliency-Based Explainable artificial intelligence (XAI) methods in Deep Learning-Based mammogram analysis.基于显著性的可解释人工智能(XAI)方法在基于深度学习的乳房X光片分析中的定量评估。
Eur J Radiol. 2024 Apr;173:111356. doi: 10.1016/j.ejrad.2024.111356. Epub 2024 Feb 5.
10
Explainability of deep neural networks for MRI analysis of brain tumors.深度神经网络在脑肿瘤 MRI 分析中的可解释性。
Int J Comput Assist Radiol Surg. 2022 Sep;17(9):1673-1683. doi: 10.1007/s11548-022-02619-x. Epub 2022 Apr 23.

引用本文的文献

1
Machine Learning Driven by Magnetic Resonance Imaging for the Classification of Alzheimer Disease Progression: Systematic Review and Meta-Analysis.磁共振成像驱动的机器学习用于阿尔茨海默病进展分类:系统评价与荟萃分析
JMIR Aging. 2024 Dec 23;7:e59370. doi: 10.2196/59370.
2
Artificial intelligence-based prediction of pathogen emergence and evolution in the world of synthetic biology.基于人工智能的预测病原体在合成生物学世界中的出现和进化。
Microb Biotechnol. 2024 Oct;17(10):e70014. doi: 10.1111/1751-7915.70014.
3
XAI-TRIS: non-linear image benchmarks to quantify false positive post-hoc attribution of feature importance.

本文引用的文献

1
XAI-TRIS: non-linear image benchmarks to quantify false positive post-hoc attribution of feature importance.XAI-TRIS:用于量化特征重要性的事后误报归因的非线性图像基准。
Mach Learn. 2024;113(9):6871-6910. doi: 10.1007/s10994-024-06574-3. Epub 2024 Jul 16.
2
Towards the interpretability of deep learning models for multi-modal neuroimaging: Finding structural changes of the ageing brain.迈向多模态神经影像学深度学习模型的可解释性:寻找大脑老化的结构变化。
Neuroimage. 2022 Nov 1;261:119504. doi: 10.1016/j.neuroimage.2022.119504. Epub 2022 Jul 23.
3
Scrutinizing XAI using linear ground-truth data with suppressor variables.
XAI-TRIS:用于量化特征重要性的事后误报归因的非线性图像基准。
Mach Learn. 2024;113(9):6871-6910. doi: 10.1007/s10994-024-06574-3. Epub 2024 Jul 16.
使用带有抑制变量的线性真实数据来审视可解释人工智能。
Mach Learn. 2022;111(5):1903-1923. doi: 10.1007/s10994-022-06167-y. Epub 2022 Apr 13.
4
Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead.停止为高风险决策解释黑箱机器学习模型,转而使用可解释模型。
Nat Mach Intell. 2019 May;1(5):206-215. doi: 10.1038/s42256-019-0048-x. Epub 2019 May 13.
5
Transfer Learning Approaches for Neuroimaging Analysis: A Scoping Review.神经影像学分析的迁移学习方法:一项范围综述
Front Artif Intell. 2022 Feb 21;5:780405. doi: 10.3389/frai.2022.780405. eCollection 2022.
6
Transfer Learning in Magnetic Resonance Brain Imaging: A Systematic Review.磁共振脑成像中的迁移学习:系统综述
J Imaging. 2021 Apr 1;7(4):66. doi: 10.3390/jimaging7040066.
7
A Transfer-Learning Approach for Accelerated MRI Using Deep Neural Networks.一种使用深度神经网络加速磁共振成像的迁移学习方法。
Magn Reson Med. 2020 Aug;84(2):663-685. doi: 10.1002/mrm.28148. Epub 2020 Jan 3.
8
White matter hyperintensities are common in midlife and already associated with cognitive decline.脑白质高信号在中年人群中很常见,并且已经与认知能力下降相关联。
Brain Commun. 2019;1(1):fcz041. doi: 10.1093/braincomms/fcz041. Epub 2019 Dec 9.
9
High-performance medicine: the convergence of human and artificial intelligence.高性能医学:人机智能融合。
Nat Med. 2019 Jan;25(1):44-56. doi: 10.1038/s41591-018-0300-7. Epub 2019 Jan 7.
10
Transfer Learning with Convolutional Neural Networks for Classification of Abdominal Ultrasound Images.用于腹部超声图像分类的卷积神经网络迁移学习
J Digit Imaging. 2017 Apr;30(2):234-243. doi: 10.1007/s10278-016-9929-2.