Electrical and Computer Engineering, University of Louisville, Louisville, KY, 40292, USA.
Electrical and Computer Engineering, University of Louisville, Louisville, KY, 40292, USA; Information Technology Institute, University of Social Science, Łódz 90-113, Poland.
Neural Netw. 2019 Oct;118:148-158. doi: 10.1016/j.neunet.2019.04.021. Epub 2019 May 9.
This paper presents an efficient technique to reduce the inference cost of deep and/or wide convolutional neural network models by pruning redundant features (or filters). Previous studies have shown that over-sized deep neural network models tend to produce a lot of redundant features that are either shifted version of one another or are very similar and show little or no variations, thus resulting in filtering redundancy. We propose to prune these redundant features along with their related feature maps according to their relative cosine distances in the feature space, thus leading to smaller networks with reduced post-training inference computational costs and competitive performance. We empirically show on select models (VGG-16, ResNet-56, ResNet-110, and ResNet-34) and dataset (MNIST Handwritten digits, CIFAR-10, and ImageNet) that inference costs (in FLOPS) can be significantly reduced while overall performance is still competitive with the state-of-the-art.
本文提出了一种有效的技术,通过剪枝冗余特征(或滤波器)来降低深度和/或宽卷积神经网络模型的推理成本。先前的研究表明,过大的深度神经网络模型往往会产生大量冗余特征,这些特征要么是彼此的移位版本,要么非常相似,几乎没有或没有变化,从而导致过滤冗余。我们建议根据特征空间中的相对余弦距离,沿着它们的相关特征图来剪枝这些冗余特征,从而得到更小的网络,在减少训练后推理计算成本的同时,仍能保持具有竞争力的性能。我们在选定的模型(VGG-16、ResNet-56、ResNet-110 和 ResNet-34)和数据集(MNIST 手写数字、CIFAR-10 和 ImageNet)上进行了实验,结果表明,推理成本(以 FLOPS 为单位)可以显著降低,而整体性能仍与最先进的技术具有竞争力。