Department of Radiology, University of Michigan, Ann Arbor, MI 48109-5842, United States of America.
Phys Med Biol. 2020 May 11;65(10):105002. doi: 10.1088/1361-6560/ab82e8.
Deep convolutional neural network (DCNN), now popularly called artificial intelligence (AI), has shown the potential to improve over previous computer-assisted tools in medical imaging developed in the past decades. A DCNN has millions of free parameters that need to be trained, but the training sample set is limited in size for most medical imaging tasks so that transfer learning is typically used. Automatic data mining may be an efficient way to enlarge the collected data set but the data can be noisy such as incorrect labels or even a wrong type of image. In this work we studied the generalization error of DCNN with transfer learning in medical imaging for the task of classifying malignant and benign masses on mammograms. With a finite available data set, we simulated a training set containing corrupted data or noisy labels. The balance between learning and memorization of the DCNN was manipulated by varying the proportion of corrupted data in the training set. The generalization error of DCNN was analyzed by the area under the receiver operating characteristic curve for the training and test sets and the weight changes after transfer learning. The study demonstrates that the transfer learning strategy of DCNN for such tasks needs to be designed properly, taking into consideration the constraints of the available training set having limited size and quality for the classification task at hand, to minimize memorization and improve generalizability.
深度卷积神经网络(DCNN),现在通常被称为人工智能(AI),已经显示出在过去几十年中开发的医学影像计算机辅助工具方面取得进展的潜力。DCNN 有数百个需要训练的自由参数,但对于大多数医学影像任务来说,训练样本集的规模有限,因此通常使用迁移学习。自动数据挖掘可能是一种有效方法来扩大已收集数据集,但数据可能存在噪声,例如错误的标签,甚至是错误类型的图像。在这项工作中,我们研究了迁移学习在医学影像中的 DCNN 泛化误差,用于在乳房 X 光片中对恶性和良性肿块进行分类的任务。在有限的可用数据集上,我们模拟了包含损坏数据或嘈杂标签的训练集。通过改变训练集中损坏数据的比例来操纵 DCNN 的学习和记忆之间的平衡。通过接收者操作特征曲线下的面积分析 DCNN 的泛化误差,用于训练集和测试集,以及迁移学习后的权重变化。该研究表明,对于此类任务的 DCNN 迁移学习策略需要适当设计,考虑到手头分类任务的可用训练集的大小和质量有限的约束,以最小化记忆并提高通用性。