Geng Xu, Gao Jinxiong, Zhang Yonghui, Xu Dingtan
School of Information and Communication Engineering, Hainan University, Haikou, 570228, China.
Sci Rep. 2024 Mar 6;14(1):5570. doi: 10.1038/s41598-024-55942-5.
The increasing interest in filter pruning of convolutional neural networks stems from its inherent ability to effectively compress and accelerate these networks. Currently, filter pruning is mainly divided into two schools: norm-based and relation-based. These methods aim to selectively remove the least important filters according to predefined rules. However, the limitations of these methods lie in the inadequate consideration of filter diversity and the impact of batch normalization (BN) layers on the input of the next layer, which may lead to performance degradation. To address the above limitations of norm-based and similarity-based methods, this study conducts empirical analyses to reveal their drawbacks and subsequently introduces a groundbreaking complex hybrid weighted pruning method. By evaluating the correlations and norms between individual filters, as well as the parameters of the BN layer, our method effectively identifies and prunes the most redundant filters in a robust manner, thereby avoiding significant decreases in network performance. We conducted comprehensive and direct pruning experiments on different depths of ResNet using publicly available image classification datasets, ImageNet and CIFAR-10. The results demonstrate the significant efficacy of our approach. In particular, when applied to the ResNet-50 on the ImageNet dataset, achieves a significant reduction of 53.5% in floating-point operations, with a performance loss of only 0.6%.
对卷积神经网络滤波器剪枝的兴趣日益浓厚,源于其有效压缩和加速这些网络的内在能力。目前,滤波器剪枝主要分为两类:基于范数的和基于关系的。这些方法旨在根据预定义规则选择性地去除最不重要的滤波器。然而,这些方法的局限性在于对滤波器多样性以及批量归一化(BN)层对下一层输入的影响考虑不足,这可能导致性能下降。为了解决基于范数和基于相似度方法的上述局限性,本研究进行了实证分析以揭示其缺点,随后引入了一种开创性的复杂混合加权剪枝方法。通过评估各个滤波器之间的相关性和范数以及BN层的参数,我们的方法能够以稳健的方式有效地识别和修剪最冗余的滤波器,从而避免网络性能大幅下降。我们使用公开可用的图像分类数据集ImageNet和CIFAR-10对不同深度的ResNet进行了全面且直接的剪枝实验。结果证明了我们方法的显著有效性。特别是,当应用于ImageNet数据集上的ResNet-50时,浮点运算显著减少了53.5%,而性能损失仅为0.6%。