El-Bana Shimaa, Al-Kabbany Ahmad, Sharkas Maha
Alexandria Higher Institute of Engineering and Technology, Alexandria, Egypt.
Intelligent Systems Lab, Arab Academy for Science, Technology, and Maritime Transport, Alexandria, Egypt.
PeerJ Comput Sci. 2020 Oct 19;6:e303. doi: 10.7717/peerj-cs.303. eCollection 2020.
We are concerned with the challenge of coronavirus disease (COVID-19) detection in chest X-ray and Computed Tomography (CT) scans, and the classification and segmentation of related infection manifestations. Even though it is arguably not an established diagnostic tool, using machine learning-based analysis of COVID-19 medical scans has shown the potential to provide a preliminary digital second opinion. This can help in managing the current pandemic, and thus has been attracting significant research attention. In this research, we propose a multi-task pipeline that takes advantage of the growing advances in deep neural network models. In the first stage, we fine-tuned an Inception-v3 deep model for COVID-19 recognition using multi-modal learning, that is, using X-ray and CT scans. In addition to outperforming other deep models on the same task in the recent literature, with an attained accuracy of 99.4%, we also present comparative analysis for multi-modal learning against learning from X-ray scans alone. The second and the third stages of the proposed pipeline complement one another in dealing with different types of infection manifestations. The former features a convolutional neural network architecture for recognizing three types of manifestations, while the latter transfers learning from another knowledge domain, namely, pulmonary nodule segmentation in CT scans, to produce binary masks for segmenting the regions corresponding to these manifestations. Our proposed pipeline also features specialized streams in which multiple deep models are trained separately to segment specific types of infection manifestations, and we show the significant impact that this framework has on various performance metrics. We evaluate the proposed models on widely adopted datasets, and we demonstrate an increase of approximately 2.5% and 4.5% for dice coefficient and mean intersection-over-union (mIoU), respectively, while achieving 60% reduction in computational time, compared to the recent literature.
我们关注胸部X光和计算机断层扫描(CT)中冠状病毒病(COVID-19)检测的挑战,以及相关感染表现的分类和分割。尽管它可以说是一种尚未确立的诊断工具,但基于机器学习对COVID-19医学扫描进行分析已显示出提供初步数字第二意见的潜力。这有助于应对当前的大流行,因此一直吸引着大量研究关注。在本研究中,我们提出了一个利用深度神经网络模型不断发展的优势的多任务管道。在第一阶段,我们使用多模态学习(即使用X光和CT扫描)对Inception-v3深度模型进行微调以用于COVID-19识别。除了在近期文献中在同一任务上优于其他深度模型,达到99.4%的准确率外,我们还对多模态学习与仅从X光扫描学习进行了对比分析。所提出管道的第二和第三阶段在处理不同类型的感染表现方面相互补充。前者采用卷积神经网络架构来识别三种类型的表现,而后者从另一个知识领域(即CT扫描中的肺结节分割)迁移学习,以生成用于分割与这些表现相对应区域的二进制掩码。我们提出的管道还具有专门的流,其中多个深度模型分别进行训练以分割特定类型的感染表现,并且我们展示了该框架对各种性能指标的重大影响。我们在广泛采用的数据集上评估所提出的模型,与近期文献相比,我们证明骰子系数和平均交并比(mIoU)分别提高了约2.5%和4.5%,同时计算时间减少了60%。