• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

利用生成式人工智能研究医学影像模型和数据集。

Using generative AI to investigate medical imagery models and datasets.

机构信息

Google, Mountain View, CA, USA.

Google, Mountain View, CA, USA.

出版信息

EBioMedicine. 2024 Apr;102:105075. doi: 10.1016/j.ebiom.2024.105075. Epub 2024 Apr 1.

DOI:10.1016/j.ebiom.2024.105075
PMID:38565004
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10993140/
Abstract

BACKGROUND

AI models have shown promise in performing many medical imaging tasks. However, our ability to explain what signals these models have learned is severely lacking. Explanations are needed in order to increase the trust of doctors in AI-based models, especially in domains where AI prediction capabilities surpass those of humans. Moreover, such explanations could enable novel scientific discovery by uncovering signals in the data that aren't yet known to experts.

METHODS

In this paper, we present a workflow for generating hypotheses to understand which visual signals in images are correlated with a classification model's predictions for a given task. This approach leverages an automatic visual explanation algorithm followed by interdisciplinary expert review. We propose the following 4 steps: (i) Train a classifier to perform a given task to assess whether the imagery indeed contains signals relevant to the task; (ii) Train a StyleGAN-based image generator with an architecture that enables guidance by the classifier ("StylEx"); (iii) Automatically detect, extract, and visualize the top visual attributes that the classifier is sensitive towards. For visualization, we independently modify each of these attributes to generate counterfactual visualizations for a set of images (i.e., what the image would look like with the attribute increased or decreased); (iv) Formulate hypotheses for the underlying mechanisms, to stimulate future research. Specifically, present the discovered attributes and corresponding counterfactual visualizations to an interdisciplinary panel of experts so that hypotheses can account for social and structural determinants of health (e.g., whether the attributes correspond to known patho-physiological or socio-cultural phenomena, or could be novel discoveries).

FINDINGS

To demonstrate the broad applicability of our approach, we present results on eight prediction tasks across three medical imaging modalities-retinal fundus photographs, external eye photographs, and chest radiographs. We showcase examples where many of the automatically-learned attributes clearly capture clinically known features (e.g., types of cataract, enlarged heart), and demonstrate automatically-learned confounders that arise from factors beyond physiological mechanisms (e.g., chest X-ray underexposure is correlated with the classifier predicting abnormality, and eye makeup is correlated with the classifier predicting low hemoglobin levels). We further show that our method reveals a number of physiologically plausible, previously-unknown attributes based on the literature (e.g., differences in the fundus associated with self-reported sex, which were previously unknown).

INTERPRETATION

Our approach enables hypotheses generation via attribute visualizations and has the potential to enable researchers to better understand, improve their assessment, and extract new knowledge from AI-based models, as well as debug and design better datasets. Though not designed to infer causality, importantly, we highlight that attributes generated by our framework can capture phenomena beyond physiology or pathophysiology, reflecting the real world nature of healthcare delivery and socio-cultural factors, and hence interdisciplinary perspectives are critical in these investigations. Finally, we will release code to help researchers train their own StylEx models and analyze their predictive tasks of interest, and use the methodology presented in this paper for responsible interpretation of the revealed attributes.

FUNDING

Google.

摘要

背景

人工智能模型在执行许多医学影像任务方面表现出了很大的潜力。然而,我们解释这些模型学习到的信号的能力却严重不足。需要解释才能增加医生对基于人工智能的模型的信任,特别是在人工智能预测能力超过人类的领域。此外,这种解释可以通过揭示数据中尚未为专家所知的信号来实现新的科学发现。

方法

在本文中,我们提出了一种生成假设的工作流程,以了解图像中的哪些视觉信号与分类模型对给定任务的预测相关。这种方法利用了一种自动视觉解释算法,然后是跨学科的专家审查。我们提出了以下 4 个步骤:(i)训练一个分类器来执行给定的任务,以评估图像中是否确实包含与任务相关的信号;(ii)使用一种能够通过分类器进行指导的架构训练一个基于 StyleGAN 的图像生成器(“StylEx”);(iii)自动检测、提取和可视化分类器敏感的顶级视觉属性。对于可视化,我们独立地修改这些属性中的每一个,以生成一组图像的反事实可视化(即,增加或减少属性后的图像会是什么样子);(iv)为潜在机制形成假设,以激发未来的研究。具体来说,将发现的属性和相应的反事实可视化呈现给跨学科的专家小组,以便假设可以解释健康的社会和结构决定因素(例如,属性是否对应已知的病理生理或社会文化现象,或者是否可能是新的发现)。

结果

为了展示我们方法的广泛适用性,我们在三个医学成像模态-视网膜眼底照片、外眼照片和胸部 X 光片上展示了八个预测任务的结果。我们展示了许多自动学习的属性明显捕捉到临床已知特征的例子(例如,白内障的类型、心脏增大),并展示了自动学习到的与生理机制之外的因素有关的混杂因素(例如,X 光胸片曝光不足与分类器预测异常有关,眼部化妆与分类器预测低血红蛋白水平有关)。我们进一步表明,我们的方法根据文献揭示了一些具有生理合理性的、以前未知的属性(例如,与自我报告的性别相关的眼底差异,这是以前未知的)。

解释

我们的方法通过属性可视化来生成假设,并有可能使研究人员更好地理解、评估和从基于人工智能的模型中提取新知识,以及调试和设计更好的数据集。虽然不是为了推断因果关系,但重要的是,我们强调我们框架生成的属性可以捕捉到超越生理学或病理生理学的现象,反映出医疗保健提供和社会文化因素的现实世界性质,因此跨学科的观点在这些研究中至关重要。最后,我们将发布代码,以帮助研究人员训练自己的 StylEx 模型并分析他们感兴趣的预测任务,并使用本文中提出的方法对揭示的属性进行负责任的解释。

资金来源

谷歌。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71c9/10993140/e9344b98179e/gr1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71c9/10993140/e9344b98179e/gr1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71c9/10993140/e9344b98179e/gr1.jpg

相似文献

1
Using generative AI to investigate medical imagery models and datasets.利用生成式人工智能研究医学影像模型和数据集。
EBioMedicine. 2024 Apr;102:105075. doi: 10.1016/j.ebiom.2024.105075. Epub 2024 Apr 1.
2
Folic acid supplementation and malaria susceptibility and severity among people taking antifolate antimalarial drugs in endemic areas.在流行地区,服用抗叶酸抗疟药物的人群中,叶酸补充剂与疟疾易感性和严重程度的关系。
Cochrane Database Syst Rev. 2022 Feb 1;2(2022):CD014217. doi: 10.1002/14651858.CD014217.
3
A Machine Learning Approach with Human-AI Collaboration for Automated Classification of Patient Safety Event Reports: Algorithm Development and Validation Study.一种人机协作的机器学习方法用于患者安全事件报告的自动分类:算法开发与验证研究
JMIR Hum Factors. 2024 Jan 25;11:e53378. doi: 10.2196/53378.
4
GANterfactual-Counterfactual Explanations for Medical Non-experts Using Generative Adversarial Learning.使用生成对抗学习为医学非专业人员提供反事实-反事实解释。
Front Artif Intell. 2022 Apr 8;5:825565. doi: 10.3389/frai.2022.825565. eCollection 2022.
5
Deep convolutional neural network and IoT technology for healthcare.用于医疗保健的深度卷积神经网络和物联网技术。
Digit Health. 2024 Jan 17;10:20552076231220123. doi: 10.1177/20552076231220123. eCollection 2024 Jan-Dec.
6
Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents.人工智能算法与放射科住院医师对胸部 X 线片解读的比较。
JAMA Netw Open. 2020 Oct 1;3(10):e2022779. doi: 10.1001/jamanetworkopen.2020.22779.
7
ExAID: A multimodal explanation framework for computer-aided diagnosis of skin lesions.EXAID:一种用于皮肤损伤计算机辅助诊断的多模态解释框架。
Comput Methods Programs Biomed. 2022 Mar;215:106620. doi: 10.1016/j.cmpb.2022.106620. Epub 2022 Jan 5.
8
Explaining the black-box smoothly-A counterfactual approach.黑盒解释的平滑化——反事实方法。
Med Image Anal. 2023 Feb;84:102721. doi: 10.1016/j.media.2022.102721. Epub 2022 Dec 13.
9
Clinical Explainability Failure (CEF) & Explainability Failure Ratio (EFR) - Changing the Way We Validate Classification Algorithms.临床可解释性失败(CEF)与可解释性失败比(EFR)——改变我们验证分类算法的方式。
J Med Syst. 2022 Mar 5;46(4):20. doi: 10.1007/s10916-022-01806-2.
10
The Rise of Hypothesis-Driven Artificial Intelligence in Oncology.肿瘤学中假设驱动型人工智能的兴起
Cancers (Basel). 2024 Feb 18;16(4):822. doi: 10.3390/cancers16040822.

引用本文的文献

1
Ultrasound Imaging: Advancing the Diagnosis of Periodontal Disease.超声成像:推动牙周疾病的诊断
Dent J (Basel). 2025 Jul 29;13(8):349. doi: 10.3390/dj13080349.
2
Effectiveness and satisfaction of fully self-service fundus disease screening among middle-aged individuals: a cross-sectional study.中年人群中完全自助式眼底疾病筛查的有效性和满意度:一项横断面研究
BMJ Open Ophthalmol. 2025 Aug 26;10(1):e001950. doi: 10.1136/bmjophth-2024-001950.
3
DREAM: A framework for discovering mechanisms underlying AI prediction of protected attributes.
DREAM:一种用于发现人工智能对受保护属性预测背后机制的框架。
medRxiv. 2025 Jul 21:2024.04.09.24305289. doi: 10.1101/2024.04.09.24305289.
4
Fine-tuning LLM hyperparameters to align semantic and physiological contexts of aging-related pathways.微调大语言模型超参数以匹配衰老相关通路的语义和生理背景。
Mol Divers. 2025 Jun 6. doi: 10.1007/s11030-025-11226-2.
5
AI-Powered Object Detection in Radiology: Current Models, Challenges, and Future Direction.放射学中的人工智能驱动目标检测:当前模型、挑战与未来方向。
J Imaging. 2025 Apr 30;11(5):141. doi: 10.3390/jimaging11050141.
6
Data augmented lung cancer prediction framework using the nested case control NLST cohort.使用巢式病例对照NLST队列的数据增强肺癌预测框架。
Front Oncol. 2025 Feb 25;15:1492758. doi: 10.3389/fonc.2025.1492758. eCollection 2025.
7
Application of Generative Artificial Intelligence in Dyslipidemia Care.生成式人工智能在血脂异常护理中的应用。
J Lipid Atheroscler. 2025 Jan;14(1):77-93. doi: 10.12997/jla.2025.14.1.77. Epub 2024 Dec 10.
8
Exploring prospects, hurdles, and road ahead for generative artificial intelligence in orthopedic education and training.探索生成式人工智能在骨科教育与培训中的前景、障碍及未来之路。
BMC Med Educ. 2024 Dec 28;24(1):1544. doi: 10.1186/s12909-024-06592-8.
9
Counterfactual Diffusion Models for Mechanistic Explainability of Artificial Intelligence Models in Pathology.用于病理学中人工智能模型机制可解释性的反事实扩散模型
bioRxiv. 2025 Jan 8:2024.10.29.620913. doi: 10.1101/2024.10.29.620913.
10
AI predictive modeling of survival outcomes for renal cancer patients undergoing targeted therapy.人工智能预测模型在接受靶向治疗的肾癌患者生存结局中的应用。
Sci Rep. 2024 Oct 30;14(1):26156. doi: 10.1038/s41598-024-77638-6.