• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

无需剪枝的高效层压缩

Efficient Layer Compression Without Pruning.

作者信息

Wu Jie, Zhu Dingshun, Fang Leyuan, Deng Yue, Zhong Zhun

出版信息

IEEE Trans Image Process. 2023;32:4689-4700. doi: 10.1109/TIP.2023.3302519. Epub 2023 Aug 16.

DOI:10.1109/TIP.2023.3302519
PMID:37561618
Abstract

Network pruning is one of the chief means for improving the computational efficiency of Deep Neural Networks (DNNs). Pruning-based methods generally discard network kernels, channels, or layers, which however inevitably will disrupt original well-learned network correlation and thus lead to performance degeneration. In this work, we propose an Efficient Layer Compression (ELC) approach to efficiently compress serial layers by decoupling and merging rather than pruning. Specifically, we first propose a novel decoupling module to decouple the layers, enabling us readily merge serial layers that include both nonlinear and convolutional layers. Then, the decoupled network is losslessly merged based on the equivalent conversion of the parameters. In this way, our ELC can effectively reduce the depth of the network without destroying the correlation of the convolutional layers. To our best knowledge, we are the first to exploit the mergeability of serial convolutional layers for lossless network layer compression. Experimental results conducted on two datasets demonstrate that our method retains superior performance with a FLOPs reduction of 74.1% for VGG-16 and 54.6% for ResNet-56, respectively. In addition, our ELC improves the inference speed by 2× on Jetson AGX Xavier edge device.

摘要

网络剪枝是提高深度神经网络(DNN)计算效率的主要手段之一。基于剪枝的方法通常会丢弃网络内核、通道或层,然而这不可避免地会破坏原本学习良好的网络相关性,从而导致性能退化。在这项工作中,我们提出了一种高效层压缩(ELC)方法,通过解耦和合并而不是剪枝来有效地压缩串行层。具体来说,我们首先提出了一种新颖的解耦模块来解耦层,使我们能够轻松合并包括非线性层和卷积层的串行层。然后,基于参数的等效转换对解耦后的网络进行无损合并。通过这种方式,我们的ELC可以有效降低网络深度,而不会破坏卷积层的相关性。据我们所知,我们是首个利用串行卷积层的可合并性进行无损网络层压缩的。在两个数据集上进行的实验结果表明,我们的方法分别在VGG - 16上实现了74.1%的浮点运算次数(FLOPs)减少,在ResNet - 56上实现了54.6%的FLOPs减少,同时保持了卓越的性能。此外,我们的ELC在Jetson AGX Xavier边缘设备上使推理速度提高了2倍。

相似文献

1
Efficient Layer Compression Without Pruning.无需剪枝的高效层压缩
IEEE Trans Image Process. 2023;32:4689-4700. doi: 10.1109/TIP.2023.3302519. Epub 2023 Aug 16.
2
Weak sub-network pruning for strong and efficient neural networks.弱子网络剪枝技术:构建强大而高效的神经网络
Neural Netw. 2021 Dec;144:614-626. doi: 10.1016/j.neunet.2021.09.015. Epub 2021 Sep 30.
3
Dynamical Conventional Neural Network Channel Pruning by Genetic Wavelet Channel Search for Image Classification.基于遗传小波通道搜索的动态传统神经网络通道剪枝用于图像分类
Front Comput Neurosci. 2021 Oct 27;15:760554. doi: 10.3389/fncom.2021.760554. eCollection 2021.
4
Discrimination-Aware Network Pruning for Deep Model Compression.面向深度模型压缩的歧视感知网络剪枝。
IEEE Trans Pattern Anal Mach Intell. 2022 Aug;44(8):4035-4051. doi: 10.1109/TPAMI.2021.3066410. Epub 2022 Jul 1.
5
EDP: An Efficient Decomposition and Pruning Scheme for Convolutional Neural Network Compression.EDP:一种用于卷积神经网络压缩的高效分解与剪枝方案。
IEEE Trans Neural Netw Learn Syst. 2021 Oct;32(10):4499-4513. doi: 10.1109/TNNLS.2020.3018177. Epub 2021 Oct 5.
6
Model pruning based on filter similarity for edge device deployment.基于滤波器相似度的模型剪枝用于边缘设备部署。
Front Neurorobot. 2023 Mar 2;17:1132679. doi: 10.3389/fnbot.2023.1132679. eCollection 2023.
7
Where to Prune: Using LSTM to Guide Data-Dependent Soft Pruning.修剪位置:使用长短期记忆网络指导数据依赖的软修剪
IEEE Trans Image Process. 2021;30:293-304. doi: 10.1109/TIP.2020.3035028. Epub 2020 Nov 24.
8
A Hardware-Friendly High-Precision CNN Pruning Method and Its FPGA Implementation.一种硬件友好的高精度 CNN 剪枝方法及其 FPGA 实现。
Sensors (Basel). 2023 Jan 11;23(2):824. doi: 10.3390/s23020824.
9
Redundant feature pruning for accelerated inference in deep neural networks.冗余特征剪枝在深度神经网络中的加速推理。
Neural Netw. 2019 Oct;118:148-158. doi: 10.1016/j.neunet.2019.04.021. Epub 2019 May 9.
10
SAAF: Self-Adaptive Attention Factor-Based Taylor-Pruning on Convolutional Neural Networks.SAAF:基于自适应注意力因子的卷积神经网络泰勒剪枝法
IEEE Trans Neural Netw Learn Syst. 2025 May;36(5):8540-8553. doi: 10.1109/TNNLS.2024.3439435. Epub 2025 May 2.

引用本文的文献

1
Evaluating deep learning techniques for identifying tongue features in subthreshold depression: a prospective observational study.评估深度学习技术用于识别阈下抑郁症患者的舌部特征:一项前瞻性观察研究。
Front Psychiatry. 2024 Aug 8;15:1361177. doi: 10.3389/fpsyt.2024.1361177. eCollection 2024.