Draelos Rachel Lea, Carin Lawrence
Duke University Department of Computer Science, 308 Research Drive, Durham, NC 27705, United States of America.
Duke University Department of Electrical and Computer Engineering, Box 90291, Durham, NC 27708, United States of America.
Artif Intell Med. 2022 Oct;132:102372. doi: 10.1016/j.artmed.2022.102372. Epub 2022 Aug 12.
Understanding model predictions is critical in healthcare, to facilitate rapid verification of model correctness and to guard against use of models that exploit confounding variables. We introduce the challenging new task of explainable multiple abnormality classification in volumetric medical images, in which a model must indicate the regions used to predict each abnormality. To solve this task, we propose a multiple instance learning convolutional neural network, AxialNet, that allows identification of top slices for each abnormality. Next we incorporate HiResCAM, an attention mechanism, to identify sub-slice regions. We prove that for AxialNet, HiResCAM explanations are guaranteed to reflect the locations the model used, unlike Grad-CAM which sometimes highlights irrelevant locations. Armed with a model that produces faithful explanations, we then aim to improve the model's learning through a novel mask loss that leverages HiResCAM and 3D allowed regions to encourage the model to predict abnormalities based only on the organs in which those abnormalities appear. The 3D allowed regions are obtained automatically through a new approach, PARTITION, that combines location information extracted from radiology reports with organ segmentation maps obtained through morphological image processing. Overall, we propose the first model for explainable multi-abnormality prediction in volumetric medical images, and then use the mask loss to achieve a 33% improvement in organ localization of multiple abnormalities in the RAD-ChestCT dataset of 36,316 scans, representing the state of the art. This work advances the clinical applicability of multiple abnormality modeling in chest CT volumes.
在医疗保健领域,理解模型预测至关重要,这有助于快速验证模型的正确性,并防范使用利用混杂变量的模型。我们引入了在体积医学图像中进行可解释的多异常分类这一具有挑战性的新任务,其中模型必须指出用于预测每个异常的区域。为了解决这个任务,我们提出了一种多实例学习卷积神经网络AxialNet,它能够识别每个异常的顶层切片。接下来,我们引入了一种注意力机制HiResCAM,以识别子切片区域。我们证明,对于AxialNet,HiResCAM解释能够保证反映模型所使用的位置,这与Grad-CAM不同,后者有时会突出显示不相关的位置。有了一个能够产生可靠解释的模型后,我们旨在通过一种新颖的掩码损失来改进模型的学习,该损失利用HiResCAM和3D允许区域来鼓励模型仅基于出现异常的器官来预测异常。3D允许区域是通过一种新方法PARTITION自动获得的,该方法将从放射学报告中提取的位置信息与通过形态图像处理获得的器官分割图相结合。总体而言,我们提出了第一个用于体积医学图像中可解释多异常预测的模型,然后使用掩码损失在包含36316次扫描的RAD-ChestCT数据集中将多个异常的器官定位提高了33%,代表了当前的技术水平。这项工作推动了胸部CT体积中多异常建模在临床中的应用。