Department of Electrical and Computer Engineering, Isfahan University of Technology, 84156-8311, Iran.
Computer Science Department, Seattle University, Seattle 98122, USA.
Artif Intell Med. 2021 Nov;121:102176. doi: 10.1016/j.artmed.2021.102176. Epub 2021 Sep 17.
Over the last decade, advances in Machine Learning and Artificial Intelligence have highlighted their potential as a diagnostic tool in the healthcare domain. Despite the widespread availability of medical images, their usefulness is severely hampered by a lack of access to labeled data. For example, while Convolutional Neural Networks (CNNs) have emerged as an essential analytical tool in image processing, their impact is curtailed by training limitations due to insufficient labeled data availability. Transfer Learning enables models developed for one task to be reused for a second task. Knowledge distillation enables transferring knowledge from a pre-trained model to another. However, it suffers from limitations, and the two models' constraints need to be architecturally similar. Knowledge distillation addresses some of the shortcomings of transfer learning by generalizing a complex model to a lighter model. However, some parts of the knowledge may not be distilled by knowledge distillation sufficiently. In this paper, a novel knowledge distillation approach using transfer learning is proposed. The proposed approach transfers the complete knowledge of a model to a new smaller one. Unlabeled data are used in an unsupervised manner to transfer the new smaller model's maximum amount of knowledge. The proposed method can be beneficial in medical image analysis, where labeled data are typically scarce. The proposed approach is evaluated in classifying images for diagnosing Diabetic Retinopathy on two publicly available datasets, including Messidor and EyePACS. Simulation results demonstrate that the approach effectively transfers knowledge from a complex model to a lighter one. Furthermore, experimental results illustrate that different small models' performance is improved significantly using unlabeled data and knowledge distillation.
在过去的十年中,机器学习和人工智能的进步凸显了它们作为医疗保健领域诊断工具的潜力。尽管医疗图像广泛可用,但由于缺乏标记数据,它们的用途受到严重限制。例如,虽然卷积神经网络 (CNN) 已成为图像处理的重要分析工具,但由于缺乏标记数据,其训练受到限制,因此其影响受到限制。迁移学习使为一个任务开发的模型可用于第二个任务。知识蒸馏可以将来自预训练模型的知识转移到另一个模型中。但是,它存在局限性,并且两个模型的约束条件需要在架构上相似。知识蒸馏通过将复杂模型泛化到较轻的模型来解决迁移学习的一些缺点。然而,知识蒸馏可能无法充分提炼某些部分的知识。在本文中,提出了一种使用迁移学习的新的知识蒸馏方法。该方法将模型的完整知识转移到新的较小模型中。使用无监督的方式使用未标记的数据来转移新的较小模型的最大知识量。该方法在医学图像分析中可能很有用,因为在医学图像分析中,标记数据通常很少。在两个公开可用的数据集上,包括 Messidor 和 EyePACS,对所提出的方法进行了分类图像以诊断糖尿病视网膜病变的评估。模拟结果表明,该方法可以有效地将知识从复杂模型转移到较轻的模型。此外,实验结果表明,使用未标记的数据和知识蒸馏可以显著提高不同小模型的性能。
Artif Intell Med. 2021-11
Artif Intell Med. 2021-9
JAMA Ophthalmol. 2023-11-1
Med Biol Eng Comput. 2023-9
Cancers (Basel). 2021-3-30
Quant Imaging Med Surg. 2024-7-1
J Med Imaging (Bellingham). 2022-9