Mohammed Ahmed, Wang Congcong, Zhao Meng, Ullah Mohib, Naseem Rabia, Wang Hao, Pedersen Marius, Cheikh Faouzi Alaya
Department of Computer ScienceNorwegian University of Science and Technology (NTNU) 2815 Gjøvik Norway.
IEEE Access. 2020 Aug 21;8:155987-156000. doi: 10.1109/ACCESS.2020.3018498. eCollection 2020.
Deep Learning-based chest Computed Tomography (CT) analysis has been proven to be effective and efficient for COVID-19 diagnosis. Existing deep learning approaches heavily rely on large labeled data sets, which are difficult to acquire in this pandemic situation. Therefore, weakly-supervised approaches are in demand. In this paper, we propose an end-to-end weakly-supervised COVID-19 detection approach, ResNext+, that only requires volume level data labels and can provide slice level prediction. The proposed approach incorporates a lung segmentation mask as well as spatial and channel attention to extract spatial features. Besides, Long Short Term Memory (LSTM) is utilized to acquire the axial dependency of the slices. Moreover, a slice attention module is applied before the final fully connected layer to generate the slice level prediction without additional supervision. An ablation study is conducted to show the efficiency of the attention blocks and the segmentation mask block. Experimental results, obtained from publicly available datasets, show a precision of 81.9% and F1 score of 81.4%. The closest state-of-the-art gives 76.7% precision and 78.8% F1 score. The 5% improvement in precision and 3% in the F1 score demonstrate the effectiveness of the proposed method. It is worth noticing that, applying image enhancement approaches do not improve the performance of the proposed method, sometimes even harm the scores, although the enhanced images have better perceptual quality.
基于深度学习的胸部计算机断层扫描(CT)分析已被证明在新冠病毒疾病(COVID-19)诊断中有效且高效。现有的深度学习方法严重依赖大量带标注的数据集,而在这种疫情情况下很难获取。因此,弱监督方法成为需求。在本文中,我们提出了一种端到端的弱监督COVID-19检测方法ResNext+,该方法仅需要体素级数据标签,并能提供切片级预测。所提出的方法结合了肺部分割掩码以及空间和通道注意力来提取空间特征。此外,利用长短期记忆(LSTM)来获取切片的轴向依赖性。而且,在最终的全连接层之前应用切片注意力模块以生成切片级预测,无需额外监督。进行了消融研究以展示注意力块和分割掩码块的有效性。从公开可用数据集获得的实验结果显示精度为81.9%,F1分数为81.4%。最接近的现有技术精度为76.7%,F1分数为78.8%。精度提高5%,F1分数提高3%证明了所提方法的有效性。值得注意的是,应用图像增强方法并没有提高所提方法的性能,有时甚至会损害分数,尽管增强后的图像具有更好的感知质量。