• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

ℓ 正则化变换在稀疏深度神经网络学习中的应用。

Transformed ℓ regularization for learning sparse deep neural networks.

机构信息

School of Mathematical Sciences, University of Chinese Academy of Sciences, Beijing, 100049, China.

College of Information Science and Engineering, Henan University of Technology, Zhengzhou, 450001, China.

出版信息

Neural Netw. 2019 Nov;119:286-298. doi: 10.1016/j.neunet.2019.08.015. Epub 2019 Aug 27.

DOI:10.1016/j.neunet.2019.08.015
PMID:31499353
Abstract

Deep Neural Networks (DNNs) have achieved extraordinary success in numerous areas. However, DNNs often carry a large number of weight parameters, leading to the challenge of heavy memory and computation costs. Overfitting is another challenge for DNNs when the training data are insufficient. These challenges severely hinder the application of DNNs in resource-constrained platforms. In fact, many network weights are redundant and can be removed from the network without much loss of performance. In this paper, we introduce a new non-convex integrated transformed ℓ regularizer to promote sparsity for DNNs, which removes redundant connections and unnecessary neurons simultaneously. Specifically, we apply the transformed ℓ regularizer to the matrix space of network weights and utilize it to remove redundant connections. Besides, group sparsity is integrated to remove unnecessary neurons. An efficient stochastic proximal gradient algorithm is presented to solve the new model. To the best of our knowledge, this is the first work to develop a non-convex regularizer in sparse optimization based method to simultaneously promote connection-level and neuron-level sparsity for DNNs. Experiments on public datasets demonstrate the effectiveness of the proposed method.

摘要

深度神经网络 (DNN) 在许多领域取得了非凡的成功。然而,DNN 通常携带大量的权重参数,导致内存和计算成本的挑战。当训练数据不足时,过拟合是 DNN 的另一个挑战。这些挑战严重阻碍了 DNN 在资源受限平台上的应用。事实上,许多网络权重是冗余的,可以从网络中删除而不会造成太大的性能损失。在本文中,我们引入了一种新的非凸集成变换 ℓ 正则化方法来促进 DNN 的稀疏性,同时去除冗余连接和不必要的神经元。具体来说,我们将变换 ℓ 正则化应用于网络权重的矩阵空间,并利用它来去除冗余连接。此外,集成了组稀疏性以去除不必要的神经元。提出了一种有效的随机近端梯度算法来求解新模型。据我们所知,这是第一个在稀疏优化方法中开发非凸正则化的工作,旨在同时促进 DNN 的连接级和神经元级稀疏性。在公共数据集上的实验证明了所提出方法的有效性。

相似文献

1
Transformed ℓ regularization for learning sparse deep neural networks.ℓ 正则化变换在稀疏深度神经网络学习中的应用。
Neural Netw. 2019 Nov;119:286-298. doi: 10.1016/j.neunet.2019.08.015. Epub 2019 Aug 27.
2
Sparsity-control ternary weight networks.稀疏控制三进制权值网络。
Neural Netw. 2022 Jan;145:221-232. doi: 10.1016/j.neunet.2021.10.018. Epub 2021 Oct 29.
3
GXNOR-Net: Training deep neural networks with ternary weights and activations without full-precision memory under a unified discretization framework.GXNOR-Net:在统一的离散化框架下使用三进制权重和激活函数训练深度神经网络,无需全精度存储。
Neural Netw. 2018 Apr;100:49-58. doi: 10.1016/j.neunet.2018.01.010. Epub 2018 Feb 2.
4
Compressing Deep Networks by Neuron Agglomerative Clustering.通过神经元聚合聚类压缩深度网络
Sensors (Basel). 2020 Oct 23;20(21):6033. doi: 10.3390/s20216033.
5
Deep Sparse Learning for Automatic Modulation Classification Using Recurrent Neural Networks.基于递归神经网络的自动调制分类的深度稀疏学习。
Sensors (Basel). 2021 Sep 25;21(19):6410. doi: 10.3390/s21196410.
6
Nonconvex Sparse Regularization for Deep Neural Networks and Its Optimality.非凸稀疏正则化在深度神经网络中的应用及其最优性。
Neural Comput. 2022 Jan 14;34(2):476-517. doi: 10.1162/neco_a_01457.
7
Perturbation of deep autoencoder weights for model compression and classification of tabular data.扰动深度自动编码器权重以进行模型压缩和表格数据分类。
Neural Netw. 2022 Dec;156:160-169. doi: 10.1016/j.neunet.2022.09.020. Epub 2022 Sep 27.
8
SeReNe: Sensitivity-Based Regularization of Neurons for Structured Sparsity in Neural Networks.SeReNe:基于灵敏度的神经网络神经元正则化以实现结构稀疏性
IEEE Trans Neural Netw Learn Syst. 2022 Dec;33(12):7237-7250. doi: 10.1109/TNNLS.2021.3084527. Epub 2022 Nov 30.
9
Feature flow regularization: Improving structured sparsity in deep neural networks.特征流正则化:改善深度神经网络中的结构化稀疏性。
Neural Netw. 2023 Apr;161:598-613. doi: 10.1016/j.neunet.2023.02.013. Epub 2023 Feb 13.
10
Efficient construction of sparse radial basis function neural networks using L-regularization.利用 L-正则化高效构建稀疏径向基函数神经网络。
Neural Netw. 2017 Oct;94:239-254. doi: 10.1016/j.neunet.2017.07.004. Epub 2017 Jul 27.

引用本文的文献

1
Infusing structural assumptions into dimensionality reduction for single-cell RNA sequencing data to identify small gene sets.将结构假设融入单细胞RNA测序数据的降维过程中以识别小基因集。
Commun Biol. 2025 Mar 11;8(1):414. doi: 10.1038/s42003-025-07872-9.
2
Efficient federated learning for distributed neuroimaging data.用于分布式神经影像数据的高效联邦学习
Front Neuroinform. 2024 Sep 9;18:1430987. doi: 10.3389/fninf.2024.1430987. eCollection 2024.
3
Aligned deep neural network for integrative analysis with high-dimensional input.
用于高维输入的综合分析的对齐深度神经网络。
J Biomed Inform. 2023 Aug;144:104434. doi: 10.1016/j.jbi.2023.104434. Epub 2023 Jun 28.
4
Machine learning algorithms for identifying predictive variables of mortality risk following dementia diagnosis: a longitudinal cohort study.机器学习算法在识别痴呆症诊断后死亡风险预测变量中的应用:一项纵向队列研究。
Sci Rep. 2023 Jun 10;13(1):9480. doi: 10.1038/s41598-023-36362-3.
5
Consistent Sparse Deep Learning: Theory and Computation.一致稀疏深度学习:理论与计算
J Am Stat Assoc. 2022;117(540):1981-1995. doi: 10.1080/01621459.2021.1895175. Epub 2021 Apr 20.
6
A phase transition for finding needles in nonlinear haystacks with LASSO artificial neural networks.使用套索人工神经网络在非线性干草堆中寻找针的相变。
Stat Comput. 2022;32(6):99. doi: 10.1007/s11222-022-10169-0. Epub 2022 Oct 22.
7
Survival Analysis with High-Dimensional Omics Data Using a Threshold Gradient Descent Regularization-Based Neural Network Approach.基于阈值梯度下降正则化的神经网络方法对高维组学数据进行生存分析。
Genes (Basel). 2022 Sep 19;13(9):1674. doi: 10.3390/genes13091674.