Suppr超能文献

冗余特征剪枝在深度神经网络中的加速推理。

Redundant feature pruning for accelerated inference in deep neural networks.

机构信息

Electrical and Computer Engineering, University of Louisville, Louisville, KY, 40292, USA.

Electrical and Computer Engineering, University of Louisville, Louisville, KY, 40292, USA; Information Technology Institute, University of Social Science, Łódz 90-113, Poland.

出版信息

Neural Netw. 2019 Oct;118:148-158. doi: 10.1016/j.neunet.2019.04.021. Epub 2019 May 9.

Abstract

This paper presents an efficient technique to reduce the inference cost of deep and/or wide convolutional neural network models by pruning redundant features (or filters). Previous studies have shown that over-sized deep neural network models tend to produce a lot of redundant features that are either shifted version of one another or are very similar and show little or no variations, thus resulting in filtering redundancy. We propose to prune these redundant features along with their related feature maps according to their relative cosine distances in the feature space, thus leading to smaller networks with reduced post-training inference computational costs and competitive performance. We empirically show on select models (VGG-16, ResNet-56, ResNet-110, and ResNet-34) and dataset (MNIST Handwritten digits, CIFAR-10, and ImageNet) that inference costs (in FLOPS) can be significantly reduced while overall performance is still competitive with the state-of-the-art.

摘要

本文提出了一种有效的技术,通过剪枝冗余特征(或滤波器)来降低深度和/或宽卷积神经网络模型的推理成本。先前的研究表明,过大的深度神经网络模型往往会产生大量冗余特征,这些特征要么是彼此的移位版本,要么非常相似,几乎没有或没有变化,从而导致过滤冗余。我们建议根据特征空间中的相对余弦距离,沿着它们的相关特征图来剪枝这些冗余特征,从而得到更小的网络,在减少训练后推理计算成本的同时,仍能保持具有竞争力的性能。我们在选定的模型(VGG-16、ResNet-56、ResNet-110 和 ResNet-34)和数据集(MNIST 手写数字、CIFAR-10 和 ImageNet)上进行了实验,结果表明,推理成本(以 FLOPS 为单位)可以显著降低,而整体性能仍与最先进的技术具有竞争力。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验