Kuldashboy Avazov, Umirzakova Sabina, Allaberdiev Sharofiddin, Nasimov Rashid, Abdusalomov Akmalbek, Cho Young Im
Department of Computer Engineering, Gachon University Sujeong-Gu, Seongnam-Si, Gyeonggi-Do, 461-701, Republic of Korea.
Department College of Computer Science and Software Engineering, Shenzhen University, Shenzhen 518060, China.
Heliyon. 2024 Jul 14;10(14):e34376. doi: 10.1016/j.heliyon.2024.e34376. eCollection 2024 Jul 30.
This paper introduces an innovative image classification technique utilizing knowledge distillation, tailored for a lightweight model structure. The core of the approach is a modified version of the AlexNet architecture, enhanced with depthwise-separable convolution layers. A unique aspect of this work is the Teacher-Student Collaborative Knowledge Distillation (TSKD) method. Unlike conventional knowledge distillation techniques, TSKD employs a dual-layered learning strategy, where the student model learns from both the final output and the intermediate layers of the teacher model. This collaborative learning approach enables the student model to actively engage in the learning process, resulting in more efficient knowledge transfer. The paper emphasizes the model suitability for scenarios with limited computational resources. This is achieved through architectural optimizations and the introduction of specialized loss functions, which balance the trade-off between model complexity and computational efficiency. The study demonstrates that despite its lightweight nature, the model maintains high accuracy and robustness in image classification tasks. Key contributions of the paper include the innovative use of depthwise-separable convolution in AlexNet, the TSKD approach for enhanced knowledge transfer, and the development of unique loss functions. These advancements collectively contribute to the model effectiveness in environments with computational constraints, making it a valuable contribution to the field of image classification.
本文介绍了一种创新的图像分类技术,该技术利用知识蒸馏,专为轻量级模型结构量身定制。该方法的核心是AlexNet架构的改进版本,通过深度可分离卷积层进行了增强。这项工作的一个独特之处是师生协作知识蒸馏(TSKD)方法。与传统的知识蒸馏技术不同,TSKD采用双层学习策略,其中学生模型从教师模型的最终输出和中间层进行学习。这种协作学习方法使学生模型能够积极参与学习过程,从而实现更高效的知识转移。本文强调了该模型适用于计算资源有限的场景。这是通过架构优化和引入专门的损失函数来实现的,这些损失函数平衡了模型复杂性和计算效率之间的权衡。研究表明,尽管该模型具有轻量级的特性,但在图像分类任务中仍保持较高的准确性和鲁棒性。本文的主要贡献包括在AlexNet中创新性地使用深度可分离卷积、用于增强知识转移的TSKD方法以及独特损失函数的开发。这些进展共同促进了模型在计算受限环境中的有效性,使其成为图像分类领域的一项有价值的贡献。