Kang Changhee, Kang Sang-Ug
Department of Computer Science, Sangmyung University, Seoul 03016, Republic of Korea.
Sensors (Basel). 2024 May 30;24(11):3522. doi: 10.3390/s24113522.
The purpose of this paper is to propose a novel transfer learning regularization method based on knowledge distillation. Recently, transfer learning methods have been used in various fields. However, problems such as knowledge loss still occur during the process of transfer learning to a new target dataset. To solve these problems, there are various regularization methods based on knowledge distillation techniques. In this paper, we propose a transfer learning regularization method based on feature map alignment used in the field of knowledge distillation. The proposed method is composed of two attention-based submodules: self-pixel attention (SPA) and global channel attention (GCA). The self-pixel attention submodule utilizes both the feature maps of the source and target models, so that it provides an opportunity to jointly consider the features of the target and the knowledge of the source. The global channel attention submodule determines the importance of channels through all layers, unlike the existing methods that calculate these only within a single layer. Accordingly, transfer learning regularization is performed by considering both the interior of each single layer and the depth of the entire layer. Consequently, the proposed method using both of these submodules showed overall improved classification accuracy than the existing methods in classification experiments on commonly used datasets.
本文旨在提出一种基于知识蒸馏的新型迁移学习正则化方法。最近,迁移学习方法已被应用于各个领域。然而,在向新的目标数据集进行迁移学习的过程中,仍然会出现诸如知识损失等问题。为了解决这些问题,有各种基于知识蒸馏技术的正则化方法。在本文中,我们提出了一种基于知识蒸馏领域中使用的特征图对齐的迁移学习正则化方法。所提出的方法由两个基于注意力的子模块组成:自像素注意力(SPA)和全局通道注意力(GCA)。自像素注意力子模块利用源模型和目标模型的特征图,从而提供了一个共同考虑目标特征和源知识的机会。全局通道注意力子模块通过所有层来确定通道的重要性,这与现有的仅在单个层内计算这些重要性的方法不同。因此,通过同时考虑每个单层的内部情况和整个层的深度来进行迁移学习正则化。结果,在常用数据集上的分类实验中,使用这两个子模块的所提出方法比现有方法显示出整体更高的分类准确率。