Wang Shiguang, Xie Tao, Liu Haijun, Zhang Xingcheng, Cheng Jian
University of Electronic Science and Technology of China, No. 2006, Xiyuan Ave, West Hi-Tech Zone, Chengdu, 611731, China.
Harbin Institute of Technology, Harbin, 150006, China.
Neural Netw. 2024 Jun;174:106263. doi: 10.1016/j.neunet.2024.106263. Epub 2024 Mar 20.
Channel Pruning is one of the most widespread techniques used to compress deep neural networks while maintaining their performances. Currently, a typical pruning algorithm leverages neural architecture search to directly find networks with a configurable width, the key step of which is to identify representative subnet for various pruning ratios by training a supernet. However, current methods mainly follow a serial training strategy to optimize supernet, which is very time-consuming. In this work, we introduce PSE-Net, a novel parallel-subnets estimator for efficient channel pruning. Specifically, we propose a parallel-subnets training algorithm that simulate the forward-backward pass of multiple subnets by droping extraneous features on batch dimension, thus various subnets could be trained in one round. Our proposed algorithm facilitates the efficiency of supernet training and equips the network with the ability to interpolate the accuracy of unsampled subnets, enabling PSE-Net to effectively evaluate and rank the subnets. Over the trained supernet, we develop a prior-distributed-based sampling algorithm to boost the performance of classical evolutionary search. Such algorithm utilizes the prior information of supernet training phase to assist in the search of optimal subnets while tackling the challenge of discovering samples that satisfy resource constraints due to the long-tail distribution of network configuration. Extensive experiments demonstrate PSE-Net outperforms previous state-of-the-art channel pruning methods on the ImageNet dataset while retaining superior supernet training efficiency. For example, under 300M FLOPs constraint, our pruned MobileNetV2 achieves 75.2% Top-1 accuracy on ImageNet dataset, exceeding the original MobileNetV2 by 2.6 units while only cost 30%/16% times than BCNet/AutoAlim.
通道剪枝是在保持深度神经网络性能的同时用于压缩网络的最广泛使用的技术之一。目前,一种典型的剪枝算法利用神经架构搜索直接找到具有可配置宽度的网络,其关键步骤是通过训练一个超网络来为各种剪枝率识别代表性子网。然而,当前方法主要遵循串行训练策略来优化超网络,这非常耗时。在这项工作中,我们引入了PSE-Net,一种用于高效通道剪枝的新型并行子网估计器。具体来说,我们提出了一种并行子网训练算法,该算法通过在批次维度上丢弃无关特征来模拟多个子网的前向-反向传播,从而可以在一轮中训练各种子网。我们提出的算法提高了超网络训练的效率,并使网络具备了插值未采样子网准确率的能力,使PSE-Net能够有效地评估子网并对其进行排名。在训练好的超网络之上,我们开发了一种基于先验分布的采样算法来提高经典进化搜索的性能。这种算法利用超网络训练阶段的先验信息来辅助搜索最优子网,同时应对由于网络配置的长尾分布而发现满足资源约束的样本的挑战。大量实验表明,PSE-Net在ImageNet数据集上优于先前的最先进通道剪枝方法,同时保持了卓越的超网络训练效率。例如,在300M FLOPs的约束下,我们剪枝后的MobileNetV2在ImageNet数据集上达到了75.2%的Top-1准确率,比原始的MobileNetV2高出2.6个百分点,而训练时间仅为BCNet/AutoAlim的30%/16%。