Healthcare AI Team, National Cancer Center, 323 Ilsan-ro, Ilsandong-gu, Goyang-si, Gyeonggi-do, 10408, Republic of Korea.
Lunit Inc., 27, Teheran-ro 2-gil, Gangnam-gu, Seoul, 06241, Republic of Korea.
Sci Rep. 2021 Apr 12;11(1):7924. doi: 10.1038/s41598-021-86726-w.
Image compression is used in several clinical organizations to help address the overhead associated with medical imaging. These methods reduce file size by using a compact representation of the original image. This study aimed to analyze the impact of image compression on the performance of deep learning-based models in classifying mammograms as "malignant"-cases that lead to a cancer diagnosis and treatment-or "normal" and "benign," non-malignant cases that do not require immediate medical intervention. In this retrospective study, 9111 unique mammograms-5672 normal, 1686 benign, and 1754 malignant cases were collected from the National Cancer Center in the Republic of Korea. Image compression was applied to mammograms with compression ratios (CRs) ranging from 15 to 11 K. Convolutional neural networks (CNNs) with three convolutional layers and three fully-connected layers were trained using these images to classify a mammogram as malignant or not malignant across a range of CRs using five-fold cross-validation. Models trained on images with maximum CRs of 5 K had an average area under the receiver operating characteristic curve (AUROC) of 0.87 and area under the precision-recall curve (AUPRC) of 0.75 across the five folds and compression ratios. For images compressed with CRs of 10 K and 11 K, model performance decreased (average 0.79 in AUROC and 0.49 in AUPRC). Upon generating saliency maps that visualize the areas each model views as significant for prediction, models trained on less compressed (CR < = 5 K) images had maps encapsulating a radiologist's label, while models trained on images with higher amounts of compression had maps that missed the ground truth completely. In addition, base ResNet18 models pre-trained on ImageNet and trained using compressed mammograms did not show performance improvements over our CNN model, with AUROC and AUPRC values ranging from 0.77 to 0.87 and 0.52 to 0.71 respectively when trained and tested on images with maximum CRs of 5 K. This paper finds that while training models on images with increased the robustness of the models when tested on compressed data, moderate image compression did not substantially impact the classification performance of DL-based models.
图像压缩在许多临床机构中被用于帮助解决与医学成像相关的开销。这些方法通过使用原始图像的紧凑表示来减小文件大小。本研究旨在分析图像压缩对基于深度学习的模型在对乳腺 X 线照片进行分类(恶性病例,导致癌症诊断和治疗,或正常和良性,非恶性病例,不需要立即进行医学干预)的性能的影响。在这项回顾性研究中,从韩国国家癌症中心收集了 9111 张独特的乳腺 X 线照片-5672 张正常、1686 张良性和 1754 张恶性病例。对乳腺 X 线照片进行了压缩,压缩比(CR)范围为 15 至 11 K。使用这些图像训练了具有三个卷积层和三个全连接层的卷积神经网络(CNN),并使用五重交叉验证在一系列 CR 下对乳腺 X 线照片进行分类,判断其为恶性或非恶性。在最大 CR 为 5 K 的图像上训练的模型在五个折叠和压缩比上的平均接收器操作特征曲线下面积(AUROC)为 0.87,平均精度召回曲线下面积(AUPRC)为 0.75。对于压缩比为 10 K 和 11 K 的图像,模型性能下降(AUROC 的平均为 0.79,AUPRC 的平均为 0.49)。在生成用于可视化每个模型认为对预测重要的区域的显着性图之后,在压缩程度较低(CR ≤ 5 K)的图像上训练的模型的地图包含放射科医生的标签,而在具有较高压缩量的图像上训练的模型的地图则完全错过了真实标签。此外,在 ImageNet 上预训练的基础 ResNet18 模型,并使用压缩的乳腺 X 线照片进行训练,其性能并未优于我们的 CNN 模型,当在最大 CR 为 5 K 的图像上进行训练和测试时,AUROC 和 AUPRC 值的范围分别为 0.77 至 0.87 和 0.52 至 0.71。本文发现,虽然在增强模型的鲁棒性的图像上训练模型,但适度的图像压缩对基于深度学习的模型的分类性能没有实质性影响。