• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

带谱随机失活的深度神经网络正则化。

Regularization of deep neural networks with spectral dropout.

机构信息

Data61, Commonwealth Scientific and Industrial Research Organization (CSIRO), Canberra ACT 2601, Australia; The Australian National University, Canberra ACT 0200, Australia.

University of Canberra, Bruce ACT 2617, Australia; Inception Institute of AI, Abu Dhabi, United Arab Emirates.

出版信息

Neural Netw. 2019 Feb;110:82-90. doi: 10.1016/j.neunet.2018.09.009. Epub 2018 Oct 16.

DOI:10.1016/j.neunet.2018.09.009
PMID:30504041
Abstract

The big breakthrough on the ImageNet challenge in 2012 was partially due to the 'Dropout' technique used to avoid overfitting. Here, we introduce a new approach called 'Spectral Dropout' to improve the generalization ability of deep neural networks. We cast the proposed approach in the form of regular Convolutional Neural Network (CNN) weight layers using a decorrelation transform with fixed basis functions. Our spectral dropout method prevents overfitting by eliminating weak and 'noisy' Fourier domain coefficients of the neural network activations, leading to remarkably better results than the current regularization methods. Furthermore, the proposed is very efficient due to the fixed basis functions used for spectral transformation. In particular, compared to Dropout and Drop-Connect, our method significantly speeds up the network convergence rate during the training process (roughly ×2), with considerably higher neuron pruning rates (an increase of ∼30%). We demonstrate that the spectral dropout can also be used in conjunction with other regularization approaches resulting in additional performance gains.

摘要

2012 年的 ImageNet 挑战赛上的重大突破部分归功于用于避免过拟合的“Dropout”技术。在这里,我们引入了一种称为“谱随机失活”的新方法来提高深度神经网络的泛化能力。我们通过使用具有固定基函数的去相关变换,将所提出的方法表示为常规卷积神经网络(CNN)的权重层。我们的谱随机失活方法通过消除神经网络激活的弱和“嘈杂”的傅里叶域系数来防止过拟合,从而导致比当前正则化方法更好的结果。此外,由于使用固定的基函数进行谱变换,因此所提出的方法非常有效。特别地,与 Dropout 和 Drop-Connect 相比,我们的方法在训练过程中显著提高了网络的收敛速度(大约快 2 倍),同时具有更高的神经元修剪率(增加约 30%)。我们证明,谱随机失活也可以与其他正则化方法结合使用,从而获得额外的性能提升。

相似文献

1
Regularization of deep neural networks with spectral dropout.带谱随机失活的深度神经网络正则化。
Neural Netw. 2019 Feb;110:82-90. doi: 10.1016/j.neunet.2018.09.009. Epub 2018 Oct 16.
2
Biased Dropout and Crossmap Dropout: Learning towards effective Dropout regularization in convolutional neural network.有偏置 Dropout 和 Crossmap Dropout:在卷积神经网络中学习有效的 Dropout 正则化。
Neural Netw. 2018 Aug;104:60-67. doi: 10.1016/j.neunet.2018.03.016. Epub 2018 Apr 9.
3
Towards dropout training for convolutional neural networks.面向卷积神经网络的随机失活训练
Neural Netw. 2015 Nov;71:1-10. doi: 10.1016/j.neunet.2015.07.007. Epub 2015 Jul 29.
4
Differential convolutional neural network.差异卷积神经网络。
Neural Netw. 2019 Aug;116:279-287. doi: 10.1016/j.neunet.2019.04.025. Epub 2019 May 10.
5
EDropout: Energy-Based Dropout and Pruning of Deep Neural Networks.EDropout:基于能量的深度神经网络的随机失活和剪枝。
IEEE Trans Neural Netw Learn Syst. 2022 Oct;33(10):5279-5292. doi: 10.1109/TNNLS.2021.3069970. Epub 2022 Oct 5.
6
Forward propagation dropout in deep neural networks using Jensen-Shannon and random forest feature importance ranking.基于 Jensen-Shannon 和随机森林特征重要性排序的深度神经网络前向传播随机失活。
Neural Netw. 2023 Aug;165:238-247. doi: 10.1016/j.neunet.2023.05.044. Epub 2023 May 29.
7
Continuous Dropout.
IEEE Trans Neural Netw Learn Syst. 2018 Sep;29(9):3926-3937. doi: 10.1109/TNNLS.2017.2750679. Epub 2017 Oct 3.
8
Deep Convolutional Neural Networks for large-scale speech tasks.用于大规模语音任务的深度卷积神经网络。
Neural Netw. 2015 Apr;64:39-48. doi: 10.1016/j.neunet.2014.08.005. Epub 2014 Sep 16.
9
Equivalence between dropout and data augmentation: A mathematical check.辍学与数据增强等效性:数学检验。
Neural Netw. 2019 Jul;115:82-89. doi: 10.1016/j.neunet.2019.03.013. Epub 2019 Mar 27.
10
Theory of adaptive SVD regularization for deep neural networks.自适应 SVD 正则化的深度神经网络理论。
Neural Netw. 2020 Aug;128:33-46. doi: 10.1016/j.neunet.2020.04.021. Epub 2020 Apr 25.

引用本文的文献

1
Prediction of Cerebrospinal Fluid (CSF) Pressure with Generative Adversarial Network Synthetic Plasma-CSF Biomarker Pairing.利用生成对抗网络合成血浆-脑脊液生物标志物配对预测脑脊液(CSF)压力
Neuroinformatics. 2025 Jul 10;23(3):38. doi: 10.1007/s12021-025-09729-2.
2
Deep Learning in Digital Breast Tomosynthesis: Current Status, Challenges, and Future Trends.数字乳腺断层合成中的深度学习:现状、挑战与未来趋势。
MedComm (2020). 2025 Jun 9;6(6):e70247. doi: 10.1002/mco2.70247. eCollection 2025 Jun.
3
Effects of AST-120 on mortality in patients with chronic kidney disease modeled by artificial intelligence or traditional statistical analysis.
基于人工智能或传统统计分析建模的慢性肾病患者中 AST-120 对死亡率的影响。
Sci Rep. 2024 Jan 6;14(1):738. doi: 10.1038/s41598-024-51498-6.
4
Gene Self-Expressive Networks as a Generalization-Aware Tool to Model Gene Regulatory Networks.基因自我表达网络作为一种具有泛化意识的工具,用于模拟基因调控网络。
Biomolecules. 2023 Mar 13;13(3):526. doi: 10.3390/biom13030526.
5
Development of Machine-Learning Model to Predict COVID-19 Mortality: Application of Ensemble Model and Regarding Feature Impacts.用于预测新冠肺炎死亡率的机器学习模型的开发:集成模型的应用及特征影响分析
Diagnostics (Basel). 2022 Jun 14;12(6):1464. doi: 10.3390/diagnostics12061464.
6
Predicting Monoterpene Indole Alkaloid-Related Genes from Expression Data with Artificial Neural Networks.利用人工神经网络从表达数据中预测单萜吲哚生物碱相关基因。
Methods Mol Biol. 2022;2505:131-140. doi: 10.1007/978-1-0716-2349-7_10.
7
Prediction of Protein-Protein Interactions in , Maize, and Rice by Combining Deep Neural Network With Discrete Hilbert Transform.通过结合深度神经网络与离散希尔伯特变换预测玉米和水稻中的蛋白质-蛋白质相互作用
Front Genet. 2021 Sep 20;12:745228. doi: 10.3389/fgene.2021.745228. eCollection 2021.
8
Development of machine learning model for diagnostic disease prediction based on laboratory tests.基于实验室检查的机器学习模型在疾病诊断预测中的开发。
Sci Rep. 2021 Apr 7;11(1):7567. doi: 10.1038/s41598-021-87171-5.
9
Deep Learning in LncRNAome: Contribution, Challenges, and Perspectives.长链非编码核糖核酸组学中的深度学习:贡献、挑战与展望
Noncoding RNA. 2020 Nov 30;6(4):47. doi: 10.3390/ncrna6040047.
10
Neuromorphic Computing Using NAND Flash Memory Architecture With Pulse Width Modulation Scheme.采用脉宽调制方案的基于与非闪存架构的神经形态计算
Front Neurosci. 2020 Sep 18;14:571292. doi: 10.3389/fnins.2020.571292. eCollection 2020.