Key Laboratory of Water Big Data Technology of Ministry of Water Resources, Hohai University, Nanjing 210093, China; College of Computer and Information, Hohai University, Nanjing 210093, China.
Key Laboratory of Water Big Data Technology of Ministry of Water Resources, Hohai University, Nanjing 210093, China; College of Computer and Information, Hohai University, Nanjing 210093, China; Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130015, China.
Methods. 2023 Oct;218:110-117. doi: 10.1016/j.ymeth.2023.08.001. Epub 2023 Aug 4.
Deep learning has brought a significant progress in medical image analysis. However, their lack of interpretability might bring high risk for wrong diagnosis with limited clinical knowledge embedding. In other words, we believe it's crucial for humans to interpret how deep learning work for medical analysis, thus appropriately adding knowledge constraints to correct the bias of wrong results. With such purpose, we propose Representation Group-Disentangling Network (RGD-Net) to explain the process of feature extraction and decision making inside deep learning framework, where we completely disentangle feature space of input X-ray images into independent feature groups, and each group would contribute to diagnose of a specific disease. Specifically, we first state problem definition for interpretable prediction with auto-encoder structure. Then, group-disentangled representations are extracted from input X-ray images with the proposed Group-Disentangle Module, which constructs semantic latent space by enforcing semantic consistency of attributes. Afterwards, adversarial constricts on mapping from features to diseases are proposed to prevent model collapse during training. Finally, a novel design of local tuning medical application is proposed based on RGB-Net, which is capable to aid clinicians for reasonable diagnosis. By conducting quantity of experiments on public datasets, RGD-Net have been superior to comparative studies by leveraging potential factors contributing to different diseases. We believe our work could bring interpretability in digging inherent patterns of deep learning on medical image analysis.
深度学习在医学图像分析方面取得了重大进展。然而,由于缺乏临床知识嵌入,它们的可解释性可能会带来误诊的高风险。换句话说,我们相信对于人类来说,解释深度学习如何用于医学分析是至关重要的,从而适当地添加知识约束来纠正错误结果的偏差。基于此目的,我们提出了表示群组解缠网络(RGD-Net),以解释深度学习框架内部的特征提取和决策过程,其中我们将输入 X 射线图像的特征空间完全解缠成独立的特征群组,每个群组都有助于诊断特定的疾病。具体来说,我们首先用自动编码器结构定义了可解释预测的问题定义。然后,通过提出的群组解缠模块从输入 X 射线图像中提取群组解缠表示,该模块通过强制属性的语义一致性来构建语义潜在空间。之后,提出了对从特征到疾病的映射的对抗约束,以防止模型在训练过程中崩溃。最后,基于 RGB-Net 提出了一种新的局部调整医疗应用的设计,它能够帮助临床医生进行合理的诊断。通过在公共数据集上进行大量实验,RGD-Net 通过利用导致不同疾病的潜在因素,在可比研究中表现出色。我们相信我们的工作可以为挖掘医学图像分析中深度学习的内在模式带来可解释性。