• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

EDP:一种用于卷积神经网络压缩的高效分解与剪枝方案。

EDP: An Efficient Decomposition and Pruning Scheme for Convolutional Neural Network Compression.

作者信息

Ruan Xiaofeng, Liu Yufan, Yuan Chunfeng, Li Bing, Hu Weiming, Li Yangxi, Maybank Stephen

出版信息

IEEE Trans Neural Netw Learn Syst. 2021 Oct;32(10):4499-4513. doi: 10.1109/TNNLS.2020.3018177. Epub 2021 Oct 5.

DOI:10.1109/TNNLS.2020.3018177
PMID:33136545
Abstract

Model compression methods have become popular in recent years, which aim to alleviate the heavy load of deep neural networks (DNNs) in real-world applications. However, most of the existing compression methods have two limitations: 1) they usually adopt a cumbersome process, including pretraining, training with a sparsity constraint, pruning/decomposition, and fine-tuning. Moreover, the last three stages are usually iterated multiple times. 2) The models are pretrained under explicit sparsity or low-rank assumptions, which are difficult to guarantee wide appropriateness. In this article, we propose an efficient decomposition and pruning (EDP) scheme via constructing a compressed-aware block that can automatically minimize the rank of the weight matrix and identify the redundant channels. Specifically, we embed the compressed-aware block by decomposing one network layer into two layers: a new weight matrix layer and a coefficient matrix layer. By imposing regularizers on the coefficient matrix, the new weight matrix learns to become a low-rank basis weight, and its corresponding channels become sparse. In this way, the proposed compressed-aware block simultaneously achieves low-rank decomposition and channel pruning by only one single data-driven training stage. Moreover, the network of architecture is further compressed and optimized by a novel Pruning & Merging (PM) module which prunes redundant channels and merges redundant decomposed layers. Experimental results (17 competitors) on different data sets and networks demonstrate that the proposed EDP achieves a high compression ratio with acceptable accuracy degradation and outperforms state-of-the-arts on compression rate, accuracy, inference time, and run-time memory.

摘要

近年来,模型压缩方法变得很流行,其旨在减轻深度神经网络(DNN)在实际应用中的沉重负担。然而,现有的大多数压缩方法有两个局限性:1)它们通常采用繁琐的过程,包括预训练、带稀疏性约束的训练、剪枝/分解和微调。此外,后三个阶段通常要迭代多次。2)模型是在显式稀疏性或低秩假设下进行预训练的,这难以保证广泛的适用性。在本文中,我们通过构建一个能自动最小化权重矩阵秩并识别冗余通道的压缩感知模块,提出了一种高效分解与剪枝(EDP)方案。具体来说,我们通过将一个网络层分解为两层来嵌入压缩感知模块:一个新的权重矩阵层和一个系数矩阵层。通过对系数矩阵施加正则化器,新的权重矩阵学习成为低秩基权重,并且其相应的通道变得稀疏。这样,所提出的压缩感知模块仅通过一个单一的数据驱动训练阶段就同时实现了低秩分解和通道剪枝。此外,通过一个新颖的剪枝与合并(PM)模块对网络架构进一步进行压缩和优化,该模块会剪枝冗余通道并合并冗余分解层。在不同数据集和网络上的实验结果(与17个竞争对手比较)表明,所提出的EDP在可接受的精度下降情况下实现了高压缩率,并且在压缩率、精度、推理时间和运行时内存方面优于现有技术。

相似文献

1
EDP: An Efficient Decomposition and Pruning Scheme for Convolutional Neural Network Compression.EDP:一种用于卷积神经网络压缩的高效分解与剪枝方案。
IEEE Trans Neural Netw Learn Syst. 2021 Oct;32(10):4499-4513. doi: 10.1109/TNNLS.2020.3018177. Epub 2021 Oct 5.
2
Discrimination-Aware Network Pruning for Deep Model Compression.面向深度模型压缩的歧视感知网络剪枝。
IEEE Trans Pattern Anal Mach Intell. 2022 Aug;44(8):4035-4051. doi: 10.1109/TPAMI.2021.3066410. Epub 2022 Jul 1.
3
Weak sub-network pruning for strong and efficient neural networks.弱子网络剪枝技术:构建强大而高效的神经网络
Neural Netw. 2021 Dec;144:614-626. doi: 10.1016/j.neunet.2021.09.015. Epub 2021 Sep 30.
4
Random pruning: channel sparsity by expectation scaling factor.随机剪枝:通过期望缩放因子实现通道稀疏性
PeerJ Comput Sci. 2023 Sep 5;9:e1564. doi: 10.7717/peerj-cs.1564. eCollection 2023.
5
Toward Compact ConvNets via Structure-Sparsity Regularized Filter Pruning.通过结构稀疏正则化滤波器剪枝实现紧凑卷积神经网络
IEEE Trans Neural Netw Learn Syst. 2020 Feb;31(2):574-588. doi: 10.1109/TNNLS.2019.2906563. Epub 2019 Apr 12.
6
Efficient Layer Compression Without Pruning.无需剪枝的高效层压缩
IEEE Trans Image Process. 2023;32:4689-4700. doi: 10.1109/TIP.2023.3302519. Epub 2023 Aug 16.
7
Dynamical Conventional Neural Network Channel Pruning by Genetic Wavelet Channel Search for Image Classification.基于遗传小波通道搜索的动态传统神经网络通道剪枝用于图像分类
Front Comput Neurosci. 2021 Oct 27;15:760554. doi: 10.3389/fncom.2021.760554. eCollection 2021.
8
Three-Stage Global Channel Pruning for Resources-Limited Platform.面向资源受限平台的三阶段全局通道剪枝
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):16153-16166. doi: 10.1109/TNNLS.2023.3292152. Epub 2024 Oct 29.
9
Compact Model Training by Low-Rank Projection With Energy Transfer.基于能量转移的低秩投影紧凑模型训练
IEEE Trans Neural Netw Learn Syst. 2025 Apr;36(4):6708-6722. doi: 10.1109/TNNLS.2024.3400928. Epub 2025 Apr 4.
10
Perturbation of deep autoencoder weights for model compression and classification of tabular data.扰动深度自动编码器权重以进行模型压缩和表格数据分类。
Neural Netw. 2022 Dec;156:160-169. doi: 10.1016/j.neunet.2022.09.020. Epub 2022 Sep 27.

引用本文的文献

1
Implementation of Lightweight Convolutional Neural Networks via Layer-Wise Differentiable Compression.通过逐层可微分压缩实现轻量化卷积神经网络。
Sensors (Basel). 2021 May 16;21(10):3464. doi: 10.3390/s21103464.