Zhang Yufeng, Kohne Joseph, Wittrup Emily, Najarian Kayvan
Department of Computational Medicine and Bioinformatics, University of Michigan, Ann Arbor, MI 48109, USA.
Department of Pediatrics, University of Michigan, Ann Arbor, MI 48103, USA.
Diagnostics (Basel). 2024 Jul 29;14(15):1634. doi: 10.3390/diagnostics14151634.
Pediatric respiratory disease diagnosis and subsequent treatment require accurate and interpretable analysis. A chest X-ray is the most cost-effective and rapid method for identifying and monitoring various thoracic diseases in children. Recent developments in self-supervised and transfer learning have shown their potential in medical imaging, including chest X-ray areas. In this article, we propose a three-stage framework with knowledge transfer from adult chest X-rays to aid the diagnosis and interpretation of pediatric thorax diseases. We conducted comprehensive experiments with different pre-training and fine-tuning strategies to develop transformer or convolutional neural network models and then evaluate them qualitatively and quantitatively. The ViT-Base/16 model, fine-tuned with the CheXpert dataset, a large chest X-ray dataset, emerged as the most effective, achieving a mean AUC of 0.761 (95% CI: 0.759-0.763) across six disease categories and demonstrating a high sensitivity (average 0.639) and specificity (average 0.683), which are indicative of its strong discriminative ability. The baseline models, ViT-Small/16 and ViT-Base/16, when directly trained on the Pediatric CXR dataset, only achieved mean AUC scores of 0.646 (95% CI: 0.641-0.651) and 0.654 (95% CI: 0.648-0.660), respectively. Qualitatively, our model excels in localizing diseased regions, outperforming models pre-trained on ImageNet and other fine-tuning approaches, thus providing superior explanations. The source code is available online and the data can be obtained from PhysioNet.
儿科呼吸道疾病的诊断及后续治疗需要准确且可解释的分析。胸部X光检查是识别和监测儿童各种胸部疾病最具成本效益且快速的方法。自监督学习和迁移学习的最新进展已在医学成像领域,包括胸部X光区域,展现出其潜力。在本文中,我们提出了一个三阶段框架,通过从成人胸部X光进行知识迁移,以辅助儿科胸部疾病的诊断和解读。我们采用不同的预训练和微调策略进行了全面实验,以开发Transformer或卷积神经网络模型,然后对其进行定性和定量评估。使用大型胸部X光数据集CheXpert进行微调的ViT-Base/16模型成为最有效的模型,在六个疾病类别中平均AUC达到0.761(95%置信区间:0.759 - 0.763),并显示出高敏感性(平均0.639)和特异性(平均0.683),这表明其具有很强的判别能力。基线模型ViT-Small/16和ViT-Base/16在直接在儿科CXR数据集上训练时,平均AUC分数分别仅达到0.646(95%置信区间:0.641 - 0.651)和0.654(95%置信区间:0.648 - 0.660)。定性地说,我们的模型在定位病变区域方面表现出色,优于在ImageNet上预训练的模型和其他微调方法,从而提供了更好的解释。源代码可在线获取,数据可从PhysioNet获得。