• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

Shakeout:一种正则化深度神经网络训练的新方法。

Shakeout: A New Approach to Regularized Deep Neural Network Training.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2018 May;40(5):1245-1258. doi: 10.1109/TPAMI.2017.2701831. Epub 2017 May 5.

DOI:10.1109/TPAMI.2017.2701831
PMID:28489533
Abstract

Recent years have witnessed the success of deep neural networks in dealing with a plenty of practical problems. Dropout has played an essential role in many successful deep neural networks, by inducing regularization in the model training. In this paper, we present a new regularized training approach: Shakeout. Instead of randomly discarding units as Dropout does at the training stage, Shakeout randomly chooses to enhance or reverse each unit's contribution to the next layer. This minor modification of Dropout has the statistical trait: the regularizer induced by Shakeout adaptively combines , and regularization terms. Our classification experiments with representative deep architectures on image datasets MNIST, CIFAR-10 and ImageNet show that Shakeout deals with over-fitting effectively and outperforms Dropout. We empirically demonstrate that Shakeout leads to sparser weights under both unsupervised and supervised settings. Shakeout also leads to the grouping effect of the input units in a layer. Considering the weights in reflecting the importance of connections, Shakeout is superior to Dropout, which is valuable for the deep model compression. Moreover, we demonstrate that Shakeout can effectively reduce the instability of the training process of the deep architecture.

摘要

近年来,深度神经网络在处理大量实际问题方面取得了成功。在模型训练中引入正则化,Dropout 在许多成功的深度神经网络中发挥了重要作用。本文提出了一种新的正则化训练方法:Shakeout。Shakeout 不是像 Dropout 那样在训练阶段随机丢弃单元,而是随机选择增强或反转每个单元对下一层的贡献。这种对 Dropout 的微小修改具有统计特性:Shakeout 诱导的正则化项自适应地组合了 和正则化项。我们在图像数据集 MNIST、CIFAR-10 和 ImageNet 上使用代表性的深度架构进行分类实验表明,Shakeout 有效地处理了过拟合问题,并且优于 Dropout。我们在无监督和监督设置下都经验性地证明了 Shakeout 会导致权重更加稀疏。Shakeout 还会导致层中输入单元的分组效果。考虑到权重可以反映连接的重要性,Shakeout 优于 Dropout,这对于深度模型压缩很有价值。此外,我们还证明了 Shakeout 可以有效地降低深度架构训练过程的不稳定性。

相似文献

1
Shakeout: A New Approach to Regularized Deep Neural Network Training.Shakeout:一种正则化深度神经网络训练的新方法。
IEEE Trans Pattern Anal Mach Intell. 2018 May;40(5):1245-1258. doi: 10.1109/TPAMI.2017.2701831. Epub 2017 May 5.
2
Towards dropout training for convolutional neural networks.面向卷积神经网络的随机失活训练
Neural Netw. 2015 Nov;71:1-10. doi: 10.1016/j.neunet.2015.07.007. Epub 2015 Jul 29.
3
Biased Dropout and Crossmap Dropout: Learning towards effective Dropout regularization in convolutional neural network.有偏置 Dropout 和 Crossmap Dropout:在卷积神经网络中学习有效的 Dropout 正则化。
Neural Netw. 2018 Aug;104:60-67. doi: 10.1016/j.neunet.2018.03.016. Epub 2018 Apr 9.
4
Fast DCNN based on FWT, intelligent dropout and layer skipping for image retrieval.基于 FWT、智能随机失活和层跳跃的快速 DCNN 用于图像检索。
Neural Netw. 2017 Nov;95:10-18. doi: 10.1016/j.neunet.2017.07.015. Epub 2017 Aug 8.
5
Regularization of deep neural networks with spectral dropout.带谱随机失活的深度神经网络正则化。
Neural Netw. 2019 Feb;110:82-90. doi: 10.1016/j.neunet.2018.09.009. Epub 2018 Oct 16.
6
Deep associative neural network for associative memory based on unsupervised representation learning.基于无监督表示学习的联想记忆深度联想神经网络。
Neural Netw. 2019 May;113:41-53. doi: 10.1016/j.neunet.2019.01.004. Epub 2019 Feb 1.
7
Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process.最大相关性最小冗余丢弃与信息核决定点过程。
Sensors (Basel). 2021 Mar 6;21(5):1846. doi: 10.3390/s21051846.
8
Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks.少即是多:深度神经网络的自适应可训练梯度辍学。
Sensors (Basel). 2023 Jan 24;23(3):1325. doi: 10.3390/s23031325.
9
Forward propagation dropout in deep neural networks using Jensen-Shannon and random forest feature importance ranking.基于 Jensen-Shannon 和随机森林特征重要性排序的深度神经网络前向传播随机失活。
Neural Netw. 2023 Aug;165:238-247. doi: 10.1016/j.neunet.2023.05.044. Epub 2023 May 29.
10
Supervised Deep Sparse Coding Networks for Image Classification.用于图像分类的监督深度稀疏编码网络
IEEE Trans Image Process. 2019 Jul 17. doi: 10.1109/TIP.2019.2928121.

引用本文的文献

1
Development and evaluation of a java-based deep neural network method for drug response predictions.一种基于Java的用于药物反应预测的深度神经网络方法的开发与评估。
Front Artif Intell. 2023 Mar 23;6:1069353. doi: 10.3389/frai.2023.1069353. eCollection 2023.
2
Deep insight: Convolutional neural network and its applications for COVID-19 prognosis.深度洞察:卷积神经网络及其在COVID-19预后中的应用。
Biomed Signal Process Control. 2021 Aug;69:102814. doi: 10.1016/j.bspc.2021.102814. Epub 2021 May 28.
3
Deep Neural Network With Consistency Regularization of Multi-Output Channels for Improved Tumor Detection and Delineation.
多输出通道一致性正则化的深度神经网络,用于提高肿瘤检测和勾画的性能。
IEEE Trans Med Imaging. 2021 Dec;40(12):3369-3378. doi: 10.1109/TMI.2021.3084748. Epub 2021 Nov 30.
4
A novel matrix of sequence descriptors for predicting protein-protein interactions from amino acid sequences.一种用于从氨基酸序列预测蛋白质-蛋白质相互作用的新型序列描述符矩阵。
PLoS One. 2019 Jun 7;14(6):e0217312. doi: 10.1371/journal.pone.0217312. eCollection 2019.