Department of Biomedical Informatics at the Center for Preventive Medicine and Digital Health (CPD-BW), Medical Faculty Mannheim, Heidelberg University, Theodor-Kutzer-Ufer 1-3, 68167, Mannheim, Germany.
Chair of Medical Informatics, Friedrich-Alexander-Universität Erlangen-Nürnberg, Wetterkreuz 15, 91058, Erlangen, Germany.
BMC Med Imaging. 2022 Apr 13;22(1):69. doi: 10.1186/s12880-022-00793-7.
Transfer learning (TL) with convolutional neural networks aims to improve performances on a new task by leveraging the knowledge of similar tasks learned in advance. It has made a major contribution to medical image analysis as it overcomes the data scarcity problem as well as it saves time and hardware resources. However, transfer learning has been arbitrarily configured in the majority of studies. This review paper attempts to provide guidance for selecting a model and TL approaches for the medical image classification task.
425 peer-reviewed articles were retrieved from two databases, PubMed and Web of Science, published in English, up until December 31, 2020. Articles were assessed by two independent reviewers, with the aid of a third reviewer in the case of discrepancies. We followed the PRISMA guidelines for the paper selection and 121 studies were regarded as eligible for the scope of this review. We investigated articles focused on selecting backbone models and TL approaches including feature extractor, feature extractor hybrid, fine-tuning and fine-tuning from scratch.
The majority of studies (n = 57) empirically evaluated multiple models followed by deep models (n = 33) and shallow (n = 24) models. Inception, one of the deep models, was the most employed in literature (n = 26). With respect to the TL, the majority of studies (n = 46) empirically benchmarked multiple approaches to identify the optimal configuration. The rest of the studies applied only a single approach for which feature extractor (n = 38) and fine-tuning from scratch (n = 27) were the two most favored approaches. Only a few studies applied feature extractor hybrid (n = 7) and fine-tuning (n = 3) with pretrained models.
The investigated studies demonstrated the efficacy of transfer learning despite the data scarcity. We encourage data scientists and practitioners to use deep models (e.g. ResNet or Inception) as feature extractors, which can save computational costs and time without degrading the predictive power.
利用卷积神经网络的迁移学习(TL)旨在通过利用预先学习的类似任务的知识来提高新任务的性能。它在医学图像分析中做出了重大贡献,因为它克服了数据匮乏的问题,同时节省了时间和硬件资源。然而,在大多数研究中,迁移学习被任意配置。本文试图为医学图像分类任务选择模型和 TL 方法提供指导。
从两个数据库 PubMed 和 Web of Science 中检索了 425 篇同行评议的文章,这些文章都是在 2020 年 12 月 31 日之前用英文发表的。文章由两位独立的审稿人进行评估,如果有分歧,则由第三位审稿人协助。我们遵循 PRISMA 指南进行论文选择,有 121 篇研究被认为符合本综述的范围。我们调查了专注于选择骨干模型和 TL 方法的文章,包括特征提取器、特征提取器混合、微调以及从头开始微调。
大多数研究(n=57)通过实证评估多种模型,然后是深度学习模型(n=33)和浅层模型(n=24)。在文献中,Inception 是最常用的深度模型之一(n=26)。关于 TL,大多数研究(n=46)通过实证基准测试多种方法来确定最佳配置。其余的研究只应用了一种方法,其中特征提取器(n=38)和从头开始微调(n=27)是最受欢迎的两种方法。只有少数研究应用了特征提取器混合(n=7)和微调(n=3)与预训练模型。
尽管数据匮乏,所调查的研究仍证明了迁移学习的有效性。我们鼓励数据科学家和从业者使用深度模型(例如 ResNet 或 Inception)作为特征提取器,这可以节省计算成本和时间,而不会降低预测能力。