Department of Radiology at Boston Children's Hospital, and Harvard Medical School, Boston, MA, USA.
Department of Radiology at Boston Children's Hospital, and Harvard Medical School, Boston, MA, USA.
Artif Intell Med. 2021 Jun;116:102078. doi: 10.1016/j.artmed.2021.102078. Epub 2021 Apr 23.
We present a critical assessment of the role of transfer learning in training fully convolutional networks (FCNs) for medical image segmentation. We first show that although transfer learning reduces the training time on the target task, improvements in segmentation accuracy are highly task/data-dependent. Large improvements are observed only when the segmentation task is more challenging and the target training data is smaller. We shed light on these observations by investigating the impact of transfer learning on the evolution of model parameters and learned representations. We observe that convolutional filters change little during training and still look random at convergence. We further show that quite accurate FCNs can be built by freezing the encoder section of the network at random values and only training the decoder section. At least for medical image segmentation, this finding challenges the common belief that the encoder section needs to learn data/task-specific representations. We examine the evolution of FCN representations to gain a deeper insight into the effects of transfer learning on the training dynamics. Our analysis shows that although FCNs trained via transfer learning learn different representations than FCNs trained with random initialization, the variability among FCNs trained via transfer learning can be as high as that among FCNs trained with random initialization. Moreover, feature reuse is not restricted to the early encoder layers; rather, it can be more significant in deeper layers. These findings offer new insights and suggest alternative ways of training FCNs for medical image segmentation.
我们对迁移学习在训练全卷积网络(FCN)进行医学图像分割中的作用进行了批判性评估。我们首先表明,尽管迁移学习可以减少目标任务的训练时间,但分割精度的提高高度依赖于任务/数据。只有当分割任务更具挑战性且目标训练数据较小时,才会观察到较大的提高。我们通过研究迁移学习对模型参数和学习表示的演变的影响,揭示了这些观察结果。我们发现,在训练过程中卷积滤波器变化很小,在收敛时仍然看起来很随机。我们进一步表明,通过将网络的编码器部分冻结到随机值并仅训练解码器部分,可以构建相当准确的 FCN。至少对于医学图像分割,这一发现挑战了编码器部分需要学习数据/任务特定表示的普遍观点。我们检查了 FCN 表示的演变,以更深入地了解迁移学习对训练动态的影响。我们的分析表明,尽管通过迁移学习训练的 FCN 比通过随机初始化训练的 FCN 学习不同的表示,但通过迁移学习训练的 FCN 之间的可变性可以与通过随机初始化训练的 FCN 一样高。此外,特征重用不仅限于早期的编码器层,而是在更深的层中可能更为显著。这些发现提供了新的见解,并为医学图像分割中 FCN 的训练提供了替代方法。