• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于 Jensen-Shannon 和随机森林特征重要性排序的深度神经网络前向传播随机失活。

Forward propagation dropout in deep neural networks using Jensen-Shannon and random forest feature importance ranking.

机构信息

Department of Computer Engineering, Ferdows Branch, Islamic Azad University, Ferdows, Iran.

Department of Computer Engineering, Mashhad Branch, Islamic Azad University, Mashhad, Iran.

出版信息

Neural Netw. 2023 Aug;165:238-247. doi: 10.1016/j.neunet.2023.05.044. Epub 2023 May 29.

DOI:10.1016/j.neunet.2023.05.044
PMID:37307667
Abstract

Dropout is a mechanism to prevent deep neural networks from overfitting and improving their generalization. Random dropout is the simplest method, where nodes are randomly terminated at each step of the training phase, which may lead to network accuracy reduction. In dynamic dropout, the importance of each node and its impact on the network performance is calculated, and the important nodes do not participate in the dropout. But the problem is that the importance of the nodes is not calculated consistently. A node may be considered less important and be dropped in one training epoch and on a batch of data before entering the next epoch, in which it may be an important node. On the other hand, calculating the importance of each unit in every training step is costly. In the proposed method, using random forest and Jensen-Shannon divergence, the importance of each node is calculated once. Then, in the forward propagation steps, the importance of the nodes is propagated and used in the dropout mechanism. This method is evaluated and compared with some previously proposed dropout approaches using two different deep neural network architectures on the MNIST, NorB, CIFAR10, CIFAR100, SVHN, and ImageNet datasets. The results suggest that the proposed method has better accuracy with fewer nodes and better generalizability. Also, the evaluations show that the approach has comparable complexity with other approaches and its convergence time is low as compared with state-of-the-art methods.

摘要

辍学是一种防止深度神经网络过拟合并提高其泛化能力的机制。随机辍学是最简单的方法,其中在训练阶段的每个步骤中随机终止节点,这可能导致网络精度降低。在动态辍学中,计算每个节点的重要性及其对网络性能的影响,重要节点不参与辍学。但问题是,节点的重要性不是一致计算的。一个节点在一个训练时期和在进入下一个时期之前的一批数据中可能被认为不重要而被丢弃,在这个时期中,它可能是一个重要的节点。另一方面,计算每个单元的重要性在每个训练步骤中都很昂贵。在提出的方法中,使用随机森林和 Jensen-Shannon 散度,一次计算每个节点的重要性。然后,在正向传播步骤中,节点的重要性被传播并用于辍学机制。该方法使用两种不同的深度神经网络架构在 MNIST、NorB、CIFAR10、CIFAR100、SVHN 和 ImageNet 数据集上进行了评估和比较。结果表明,该方法具有更好的准确性,更少的节点和更好的泛化能力。此外,评估表明,该方法的复杂度与其他方法相当,与最先进的方法相比,其收敛时间较低。

相似文献

1
Forward propagation dropout in deep neural networks using Jensen-Shannon and random forest feature importance ranking.基于 Jensen-Shannon 和随机森林特征重要性排序的深度神经网络前向传播随机失活。
Neural Netw. 2023 Aug;165:238-247. doi: 10.1016/j.neunet.2023.05.044. Epub 2023 May 29.
2
Regularization of deep neural networks with spectral dropout.带谱随机失活的深度神经网络正则化。
Neural Netw. 2019 Feb;110:82-90. doi: 10.1016/j.neunet.2018.09.009. Epub 2018 Oct 16.
3
Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process.最大相关性最小冗余丢弃与信息核决定点过程。
Sensors (Basel). 2021 Mar 6;21(5):1846. doi: 10.3390/s21051846.
4
Biased Dropout and Crossmap Dropout: Learning towards effective Dropout regularization in convolutional neural network.有偏置 Dropout 和 Crossmap Dropout:在卷积神经网络中学习有效的 Dropout 正则化。
Neural Netw. 2018 Aug;104:60-67. doi: 10.1016/j.neunet.2018.03.016. Epub 2018 Apr 9.
5
Hybridized sine cosine algorithm with convolutional neural networks dropout regularization application.混合正弦余弦算法与卷积神经网络辍学正则化应用。
Sci Rep. 2022 Apr 15;12(1):6302. doi: 10.1038/s41598-022-09744-2.
6
Continuous Dropout.
IEEE Trans Neural Netw Learn Syst. 2018 Sep;29(9):3926-3937. doi: 10.1109/TNNLS.2017.2750679. Epub 2017 Oct 3.
7
Towards dropout training for convolutional neural networks.面向卷积神经网络的随机失活训练
Neural Netw. 2015 Nov;71:1-10. doi: 10.1016/j.neunet.2015.07.007. Epub 2015 Jul 29.
8
Learned Gradient Compression for Distributed Deep Learning.用于分布式深度学习的学习梯度压缩
IEEE Trans Neural Netw Learn Syst. 2022 Dec;33(12):7330-7344. doi: 10.1109/TNNLS.2021.3084806. Epub 2022 Nov 30.
9
Reliable Prediction Errors for Deep Neural Networks Using Test-Time Dropout.使用测试时随机失活技术对深度神经网络进行可靠的预测误差估计。
J Chem Inf Model. 2019 Jul 22;59(7):3330-3339. doi: 10.1021/acs.jcim.9b00297. Epub 2019 Jun 26.
10
Theory of adaptive SVD regularization for deep neural networks.自适应 SVD 正则化的深度神经网络理论。
Neural Netw. 2020 Aug;128:33-46. doi: 10.1016/j.neunet.2020.04.021. Epub 2020 Apr 25.

引用本文的文献

1
Prediction and Interpretability Study of the Glass Transition Temperature of Polyimide Based on Machine Learning and Molecular Dynamics Simulations.基于机器学习和分子动力学模拟的聚酰亚胺玻璃化转变温度预测与可解释性研究
Polymers (Basel). 2025 Jul 30;17(15):2083. doi: 10.3390/polym17152083.