Huff Daniel T, Weisman Amy J, Jeraj Robert
Department of Medical Physics, University of Wisconsin-Madison, Madison WI, United States of America.
Faculty of Mathematics and Physics, University of Ljubljana, Ljubljana, Slovenia.
Phys Med Biol. 2021 Feb 2;66(4):04TR01. doi: 10.1088/1361-6560/abcd17.
Deep learning (DL) approaches to medical image analysis tasks have recently become popular; however, they suffer from a lack of human interpretability critical for both increasing understanding of the methods' operation and enabling clinical translation. This review summarizes currently available methods for performing image model interpretation and critically evaluates published uses of these methods for medical imaging applications. We divide model interpretation in two categories: (1) understanding model structure and function and (2) understanding model output. Understanding model structure and function summarizes ways to inspect the learned features of the model and how those features act on an image. We discuss techniques for reducing the dimensionality of high-dimensional data and cover autoencoders, both of which can also be leveraged for model interpretation. Understanding model output covers attribution-based methods, such as saliency maps and class activation maps, which produce heatmaps describing the importance of different parts of an image to the model prediction. We describe the mathematics behind these methods, give examples of their use in medical imaging, and compare them against one another. We summarize several published toolkits for model interpretation specific to medical imaging applications, cover limitations of current model interpretation methods, provide recommendations for DL practitioners looking to incorporate model interpretation into their task, and offer general discussion on the importance of model interpretation in medical imaging contexts.
深度学习(DL)方法在医学图像分析任务中近来颇受欢迎;然而,它们缺乏人类可解释性,而这对于增进对方法操作的理解以及实现临床转化都至关重要。本综述总结了目前用于进行图像模型解释的可用方法,并批判性地评估了这些方法在医学成像应用中的已发表用途。我们将模型解释分为两类:(1)理解模型结构和功能,以及(2)理解模型输出。理解模型结构和功能总结了检查模型学习到的特征以及这些特征如何作用于图像的方法。我们讨论了降低高维数据维度的技术,并涵盖了自动编码器,这两者都可用于模型解释。理解模型输出涵盖基于归因的方法,如显著性图和类激活图,它们生成描述图像不同部分对模型预测重要性的热图。我们描述了这些方法背后的数学原理,给出它们在医学成像中的使用示例,并相互比较。我们总结了几个专门用于医学成像应用的已发表的模型解释工具包,涵盖当前模型解释方法的局限性,为希望将模型解释纳入其任务的深度学习从业者提供建议,并对模型解释在医学成像背景下的重要性进行一般性讨论。