Hou Qinggang, Yang Wanshuai, Liu Guizhuang
School of Information Engineering, Shandong Huayu University of Technology, Dezhou, 253000, China.
School of Computer Science and Technology, Tiangong University, Tianjin, 300387, China.
Sci Rep. 2025 Jan 11;15(1):1687. doi: 10.1038/s41598-025-85697-6.
In order to reduce the number of parameters in the Chinese herbal medicine recognition model while maintaining accuracy, this paper takes 20 classes of Chinese herbs as the research object and proposes a recognition network based on knowledge distillation and cross-attention - ShuffleCANet (ShuffleNet and Cross-Attention). Firstly, transfer learning was used for experiments on 20 classic networks, and DenseNet and RegNet were selected as dual teacher models. Then, considering the parameter count and recognition accuracy, ShuffleNet was determined as the student model, and a new cross-attention mechanism was proposed. This cross-attention model replaces Conv5 in ShuffleNet to achieve the goal of lightweight design while maintaining accuracy. Finally, experiments on the public dataset NB-TCM-CHM showed that the accuracy (ACC) and F1_score of the proposed ShuffleCANet model reached 98.8%, with only 128.66M model parameters. Compared with the baseline model ShuffleNet, the parameters are reduced by nearly 50%, but the accuracy is improved by about 1.3%, proving this method's effectiveness.
为了在保持准确性的同时减少中草药识别模型中的参数数量,本文以20类中草药为研究对象,提出了一种基于知识蒸馏和交叉注意力的识别网络——ShuffleCANet(ShuffleNet和交叉注意力)。首先,对20个经典网络进行迁移学习实验,选择DenseNet和RegNet作为双教师模型。然后,考虑参数数量和识别准确率,确定ShuffleNet为学生模型,并提出了一种新的交叉注意力机制。这种交叉注意力模型取代了ShuffleNet中的Conv5,在保持准确性的同时实现了轻量化设计的目标。最后,在公共数据集NB-TCM-CHM上的实验表明,所提出的ShuffleCANet模型的准确率(ACC)和F1_score达到了98.8%,模型参数仅为128.66M。与基线模型ShuffleNet相比,参数减少了近50%,但准确率提高了约1.3%,证明了该方法的有效性。