• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过结合多种剪枝技术最小化卷积神经网络模型大小的启发式方法

Heuristic Method for Minimizing Model Size of CNN by Combining Multiple Pruning Techniques.

作者信息

Tian Danhe, Yamagiwa Shinichi, Wada Koichi

机构信息

Doctoral Program in Computer Science, University of Tsukuba, 1-1-1 Tennodai, Tsukuba 305-8573, Ibaraki, Japan.

JST, SPRING, 4-1-8 Honcho, Kawaguchi 332-0012, Saitama, Japan.

出版信息

Sensors (Basel). 2022 Aug 5;22(15):5874. doi: 10.3390/s22155874.

DOI:10.3390/s22155874
PMID:35957431
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9371397/
Abstract

Network pruning techniques have been widely used for compressing computational and memory intensive deep learning models through removing redundant components of the model. According to the pruning granularity, network pruning can be categorized into structured and unstructured methods. The structured pruning removes the large components in a model such as channels or layers, which might reduce the accuracy. The unstructured pruning directly removes mainly the parameters in a model as well as the redundant channels or layers, which might result in an inadequate pruning. To address the limitations of the pruning methods, this paper proposes a heuristic method for minimizing model size. This paper implements an algorithm to combine both the structured and the unstructured pruning methods while maintaining the target accuracy that is configured by its application. We use network slimming for the structured pruning method and deep compression for the unstructured one. Our method achieves a higher compression ratio than the case when the individual pruning method is applied. To show the effectiveness of our proposed method, this paper evaluates our proposed method with actual state-of-the-art CNN models of VGGNet, ResNet and DenseNet under the CIFAR-10 dataset. This paper discusses the performance of the proposed method with the cases of individual usage of the structured and unstructured pruning methods and then proves that our method achieves better performance with higher compression ratio. In the best case of the VGGNet, our method results in a 13× reduction ratio in the model size, and also gives a 15× reduction ratio regarding the pruning time compared with the brute-force search method.

摘要

网络剪枝技术已被广泛用于通过去除模型的冗余组件来压缩计算和内存密集型深度学习模型。根据剪枝粒度,网络剪枝可分为结构化和非结构化方法。结构化剪枝会去除模型中的大型组件,如通道或层,这可能会降低准确性。非结构化剪枝主要直接去除模型中的参数以及冗余通道或层,这可能会导致剪枝不充分。为了解决剪枝方法的局限性,本文提出了一种用于最小化模型大小的启发式方法。本文实现了一种算法,将结构化和非结构化剪枝方法相结合,同时保持由其应用配置的目标精度。我们将网络瘦身用于结构化剪枝方法,将深度压缩用于非结构化剪枝方法。我们的方法比单独应用剪枝方法时实现了更高的压缩率。为了展示我们提出的方法的有效性,本文在CIFAR-10数据集下,使用VGGNet、ResNet和DenseNet等实际的先进CNN模型对我们提出的方法进行了评估。本文讨论了结构化和非结构化剪枝方法单独使用时所提出方法的性能,然后证明了我们的方法在更高压缩率下实现了更好的性能。在VGGNet的最佳情况下,我们的方法使模型大小减少了13倍,与暴力搜索方法相比,剪枝时间也减少了15倍。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/128a7ce99069/sensors-22-05874-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/9771419b4c17/sensors-22-05874-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/41d05407ecda/sensors-22-05874-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/89ad7aeb4902/sensors-22-05874-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/5963bd8d0da0/sensors-22-05874-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/45624ff55520/sensors-22-05874-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/3f97cccefaa9/sensors-22-05874-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/81f9c01c93a7/sensors-22-05874-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/a7b59d4b506d/sensors-22-05874-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/0912b8e5be1c/sensors-22-05874-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/128a7ce99069/sensors-22-05874-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/9771419b4c17/sensors-22-05874-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/41d05407ecda/sensors-22-05874-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/89ad7aeb4902/sensors-22-05874-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/5963bd8d0da0/sensors-22-05874-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/45624ff55520/sensors-22-05874-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/3f97cccefaa9/sensors-22-05874-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/81f9c01c93a7/sensors-22-05874-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/a7b59d4b506d/sensors-22-05874-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/0912b8e5be1c/sensors-22-05874-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4b0b/9371397/128a7ce99069/sensors-22-05874-g010.jpg

相似文献

1
Heuristic Method for Minimizing Model Size of CNN by Combining Multiple Pruning Techniques.通过结合多种剪枝技术最小化卷积神经网络模型大小的启发式方法
Sensors (Basel). 2022 Aug 5;22(15):5874. doi: 10.3390/s22155874.
2
Dynamical Conventional Neural Network Channel Pruning by Genetic Wavelet Channel Search for Image Classification.基于遗传小波通道搜索的动态传统神经网络通道剪枝用于图像分类
Front Comput Neurosci. 2021 Oct 27;15:760554. doi: 10.3389/fncom.2021.760554. eCollection 2021.
3
DMPP: Differentiable multi-pruner and predictor for neural network pruning.DMPP:用于神经网络剪枝的可微分多修剪器和预测器。
Neural Netw. 2022 Mar;147:103-112. doi: 10.1016/j.neunet.2021.12.020. Epub 2021 Dec 30.
4
Random pruning: channel sparsity by expectation scaling factor.随机剪枝:通过期望缩放因子实现通道稀疏性
PeerJ Comput Sci. 2023 Sep 5;9:e1564. doi: 10.7717/peerj-cs.1564. eCollection 2023.
5
Feature flow regularization: Improving structured sparsity in deep neural networks.特征流正则化:改善深度神经网络中的结构化稀疏性。
Neural Netw. 2023 Apr;161:598-613. doi: 10.1016/j.neunet.2023.02.013. Epub 2023 Feb 13.
6
Weak sub-network pruning for strong and efficient neural networks.弱子网络剪枝技术:构建强大而高效的神经网络
Neural Netw. 2021 Dec;144:614-626. doi: 10.1016/j.neunet.2021.09.015. Epub 2021 Sep 30.
7
Where to Prune: Using LSTM to Guide Data-Dependent Soft Pruning.修剪位置:使用长短期记忆网络指导数据依赖的软修剪
IEEE Trans Image Process. 2021;30:293-304. doi: 10.1109/TIP.2020.3035028. Epub 2020 Nov 24.
8
StructADMM: Achieving Ultrahigh Efficiency in Structured Pruning for DNNs.结构化交替方向乘子法(StructADMM):在深度神经网络的结构化剪枝中实现超高效率
IEEE Trans Neural Netw Learn Syst. 2022 May;33(5):2259-2273. doi: 10.1109/TNNLS.2020.3045153. Epub 2022 May 2.
9
Deep Neural Network Compression by In-Parallel Pruning-Quantization.通过并行剪枝-量化实现深度神经网络压缩。
IEEE Trans Pattern Anal Mach Intell. 2020 Mar;42(3):568-579. doi: 10.1109/TPAMI.2018.2886192. Epub 2018 Dec 12.
10
Toward Compact ConvNets via Structure-Sparsity Regularized Filter Pruning.通过结构稀疏正则化滤波器剪枝实现紧凑卷积神经网络
IEEE Trans Neural Netw Learn Syst. 2020 Feb;31(2):574-588. doi: 10.1109/TNNLS.2019.2906563. Epub 2019 Apr 12.

本文引用的文献

1
Asymptotic Soft Filter Pruning for Deep Convolutional Neural Networks.渐进式软滤波器剪枝在深度卷积神经网络中的应用。
IEEE Trans Cybern. 2020 Aug;50(8):3594-3604. doi: 10.1109/TCYB.2019.2933477. Epub 2019 Aug 27.
2
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.