Rajaraman Sivaramakrishnan, Kim Incheol, Antani Sameer K
Lister Hill National Center for Biomedical Communications, National Library of Medicine, National Institutes of Health, Bethesda, MD, United States of America.
PeerJ. 2020 Mar 17;8:e8693. doi: 10.7717/peerj.8693. eCollection 2020.
Convolutional neural networks (CNNs) trained on natural images are extremely successful in image classification and localization due to superior automated feature extraction capability. In extending their use to biomedical recognition tasks, it is important to note that visual features of medical images tend to be uniquely different than natural images. There are advantages offered through training these networks on large scale medical common modality image collections pertaining to the recognition task. Further, improved generalization in transferring knowledge across similar tasks is possible when the models are trained to learn modality-specific features and then suitably repurposed for the target task. In this study, we propose modality-specific ensemble learning toward improving abnormality detection in chest X-rays (CXRs). CNN models are trained on a large-scale CXR collection to learn modality-specific features and then repurposed for detecting and localizing abnormalities. Model predictions are combined using different ensemble strategies toward reducing prediction variance and sensitivity to the training data while improving overall performance and generalization. Class-selective relevance mapping (CRM) is used to visualize the learned behavior of the individual models and their ensembles. It localizes discriminative regions of interest (ROIs) showing abnormal regions and offers an improved explanation of model predictions. It was observed that the model ensembles demonstrate superior localization performance in terms of Intersection of Union (IoU) and mean Average Precision (mAP) metrics than any individual constituent model.
由于具有卓越的自动特征提取能力,在自然图像上训练的卷积神经网络(CNN)在图像分类和定位方面极为成功。在将其应用扩展到生物医学识别任务时,需要注意医学图像的视觉特征往往与自然图像有着独特的差异。在与识别任务相关的大规模医学常见模态图像集上训练这些网络具有诸多优势。此外,当模型经过训练以学习特定模态特征,然后适当地重新用于目标任务时,在跨相似任务转移知识方面实现更好的泛化是可能的。在本研究中,我们提出特定模态的集成学习方法,以改进胸部X光(CXR)中的异常检测。CNN模型在大规模CXR图像集上进行训练,以学习特定模态特征,然后重新用于检测和定位异常。使用不同的集成策略组合模型预测,以减少预测方差和对训练数据的敏感性,同时提高整体性能和泛化能力。类选择性相关映射(CRM)用于可视化各个模型及其集成的学习行为。它定位显示异常区域的有区分力的感兴趣区域(ROI),并对模型预测提供更好的解释。据观察,模型集成在交并比(IoU)和平均精度均值(mAP)指标方面展示出比任何单个组成模型更优的定位性能。